mirror of
https://github.com/XRPLF/rippled.git
synced 2025-12-06 17:27:55 +00:00
Log non-account transaction in warning (RIPD-1440) Log warning on PeerImp::fail (RIPD-1444)
1350 lines
37 KiB
C++
1350 lines
37 KiB
C++
//------------------------------------------------------------------------------
|
|
/*
|
|
This file is part of rippled: https://github.com/ripple/rippled
|
|
Copyright (c) 2012, 2013 Ripple Labs Inc.
|
|
|
|
Permission to use, copy, modify, and/or distribute this software for any
|
|
purpose with or without fee is hereby granted, provided that the above
|
|
copyright notice and this permission notice appear in all copies.
|
|
|
|
THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
|
|
WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
|
|
MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
|
|
ANY SPECIAL , DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
|
|
WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
|
|
ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
|
|
OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
|
|
*/
|
|
//==============================================================================
|
|
|
|
#include <BeastConfig.h>
|
|
#include <ripple/app/ledger/Ledger.h>
|
|
#include <ripple/app/ledger/AcceptedLedger.h>
|
|
#include <ripple/app/ledger/InboundLedgers.h>
|
|
#include <ripple/app/ledger/LedgerMaster.h>
|
|
#include <ripple/app/ledger/LedgerTiming.h>
|
|
#include <ripple/app/ledger/LedgerToJson.h>
|
|
#include <ripple/app/ledger/OrderBookDB.h>
|
|
#include <ripple/app/ledger/PendingSaves.h>
|
|
#include <ripple/app/ledger/TransactionMaster.h>
|
|
#include <ripple/app/main/Application.h>
|
|
#include <ripple/app/misc/HashRouter.h>
|
|
#include <ripple/app/misc/LoadFeeTrack.h>
|
|
#include <ripple/app/misc/NetworkOPs.h>
|
|
#include <ripple/basics/contract.h>
|
|
#include <ripple/basics/Log.h>
|
|
#include <ripple/basics/StringUtilities.h>
|
|
#include <ripple/core/Config.h>
|
|
#include <ripple/core/DatabaseCon.h>
|
|
#include <ripple/core/JobQueue.h>
|
|
#include <ripple/core/SociDB.h>
|
|
#include <ripple/json/to_string.h>
|
|
#include <ripple/nodestore/Database.h>
|
|
#include <ripple/protocol/digest.h>
|
|
#include <ripple/protocol/Indexes.h>
|
|
#include <ripple/protocol/JsonFields.h>
|
|
#include <ripple/protocol/PublicKey.h>
|
|
#include <ripple/protocol/SecretKey.h>
|
|
#include <ripple/protocol/HashPrefix.h>
|
|
#include <ripple/protocol/types.h>
|
|
#include <ripple/beast/core/LexicalCast.h>
|
|
#include <boost/optional.hpp>
|
|
#include <cassert>
|
|
#include <utility>
|
|
|
|
namespace ripple {
|
|
|
|
create_genesis_t const create_genesis {};
|
|
|
|
static
|
|
uint256
|
|
calculateLedgerHash (LedgerInfo const& info)
|
|
{
|
|
// VFALCO This has to match addRaw in View.h.
|
|
return sha512Half(
|
|
HashPrefix::ledgerMaster,
|
|
std::uint32_t(info.seq),
|
|
std::uint64_t(info.drops.drops ()),
|
|
info.parentHash,
|
|
info.txHash,
|
|
info.accountHash,
|
|
std::uint32_t(info.parentCloseTime.time_since_epoch().count()),
|
|
std::uint32_t(info.closeTime.time_since_epoch().count()),
|
|
std::uint8_t(info.closeTimeResolution.count()),
|
|
std::uint8_t(info.closeFlags));
|
|
}
|
|
|
|
//------------------------------------------------------------------------------
|
|
|
|
class Ledger::sles_iter_impl
|
|
: public sles_type::iter_base
|
|
{
|
|
private:
|
|
ReadView const* view_;
|
|
SHAMap::const_iterator iter_;
|
|
|
|
public:
|
|
sles_iter_impl() = delete;
|
|
sles_iter_impl& operator= (sles_iter_impl const&) = delete;
|
|
|
|
sles_iter_impl (sles_iter_impl const&) = default;
|
|
|
|
sles_iter_impl (SHAMap::const_iterator iter,
|
|
ReadView const& view)
|
|
: view_ (&view)
|
|
, iter_ (iter)
|
|
{
|
|
}
|
|
|
|
std::unique_ptr<base_type>
|
|
copy() const override
|
|
{
|
|
return std::make_unique<
|
|
sles_iter_impl>(*this);
|
|
}
|
|
|
|
bool
|
|
equal (base_type const& impl) const override
|
|
{
|
|
auto const& other = dynamic_cast<
|
|
sles_iter_impl const&>(impl);
|
|
return iter_ == other.iter_;
|
|
}
|
|
|
|
void
|
|
increment() override
|
|
{
|
|
++iter_;
|
|
}
|
|
|
|
sles_type::value_type
|
|
dereference() const override
|
|
{
|
|
auto const item = *iter_;
|
|
SerialIter sit(item.slice());
|
|
return std::make_shared<SLE const>(
|
|
sit, item.key());
|
|
}
|
|
};
|
|
|
|
//------------------------------------------------------------------------------
|
|
|
|
class Ledger::txs_iter_impl
|
|
: public txs_type::iter_base
|
|
{
|
|
private:
|
|
bool metadata_;
|
|
ReadView const* view_;
|
|
SHAMap::const_iterator iter_;
|
|
|
|
public:
|
|
txs_iter_impl() = delete;
|
|
txs_iter_impl& operator= (txs_iter_impl const&) = delete;
|
|
|
|
txs_iter_impl (txs_iter_impl const&) = default;
|
|
|
|
txs_iter_impl (bool metadata,
|
|
SHAMap::const_iterator iter,
|
|
ReadView const& view)
|
|
: metadata_ (metadata)
|
|
, view_ (&view)
|
|
, iter_ (iter)
|
|
{
|
|
}
|
|
|
|
std::unique_ptr<base_type>
|
|
copy() const override
|
|
{
|
|
return std::make_unique<
|
|
txs_iter_impl>(*this);
|
|
}
|
|
|
|
bool
|
|
equal (base_type const& impl) const override
|
|
{
|
|
auto const& other = dynamic_cast<
|
|
txs_iter_impl const&>(impl);
|
|
return iter_ == other.iter_;
|
|
}
|
|
|
|
void
|
|
increment() override
|
|
{
|
|
++iter_;
|
|
}
|
|
|
|
txs_type::value_type
|
|
dereference() const override
|
|
{
|
|
auto const item = *iter_;
|
|
if (metadata_)
|
|
return deserializeTxPlusMeta(item);
|
|
return { deserializeTx(item), nullptr };
|
|
}
|
|
};
|
|
|
|
//------------------------------------------------------------------------------
|
|
|
|
Ledger::Ledger (
|
|
create_genesis_t,
|
|
Config const& config,
|
|
std::vector<uint256> const& amendments,
|
|
Family& family)
|
|
: mImmutable (false)
|
|
, txMap_ (std::make_shared <SHAMap> (SHAMapType::TRANSACTION,
|
|
family, SHAMap::version{1}))
|
|
, stateMap_ (std::make_shared <SHAMap> (SHAMapType::STATE,
|
|
family, SHAMap::version{1}))
|
|
, rules_{config.features}
|
|
{
|
|
info_.seq = 1;
|
|
info_.drops = SYSTEM_CURRENCY_START;
|
|
info_.closeTimeResolution = ledgerDefaultTimeResolution;
|
|
|
|
static auto const id = calcAccountID(
|
|
generateKeyPair(KeyType::secp256k1,
|
|
generateSeed("masterpassphrase")).first);
|
|
auto const sle = std::make_shared<SLE>(keylet::account(id));
|
|
sle->setFieldU32 (sfSequence, 1);
|
|
sle->setAccountID (sfAccount, id);
|
|
sle->setFieldAmount (sfBalance, info_.drops);
|
|
rawInsert(sle);
|
|
|
|
if (! amendments.empty())
|
|
{
|
|
auto const sle = std::make_shared<SLE>(keylet::amendments());
|
|
sle->setFieldV256 (sfAmendments, STVector256{amendments});
|
|
rawInsert(sle);
|
|
}
|
|
|
|
stateMap_->flushDirty (hotACCOUNT_NODE, info_.seq);
|
|
setImmutable(config);
|
|
}
|
|
|
|
Ledger::Ledger (
|
|
LedgerInfo const& info,
|
|
bool& loaded,
|
|
Config const& config,
|
|
Family& family,
|
|
beast::Journal j)
|
|
: mImmutable (true)
|
|
, txMap_ (std::make_shared <SHAMap> (SHAMapType::TRANSACTION,
|
|
info.txHash, family,
|
|
SHAMap::version{getSHAMapV2(info) ? 2 : 1}))
|
|
, stateMap_ (std::make_shared <SHAMap> (SHAMapType::STATE,
|
|
info.accountHash, family,
|
|
SHAMap::version{getSHAMapV2(info) ? 2 : 1}))
|
|
, rules_ (config.features)
|
|
, info_ (info)
|
|
{
|
|
loaded = true;
|
|
|
|
if (info_.txHash.isNonZero () &&
|
|
!txMap_->fetchRoot (SHAMapHash{info_.txHash}, nullptr))
|
|
{
|
|
loaded = false;
|
|
JLOG (j.warn()) << "Don't have TX root for ledger";
|
|
}
|
|
|
|
if (info_.accountHash.isNonZero () &&
|
|
!stateMap_->fetchRoot (SHAMapHash{info_.accountHash}, nullptr))
|
|
{
|
|
loaded = false;
|
|
JLOG (j.warn()) << "Don't have AS root for ledger";
|
|
}
|
|
|
|
txMap_->setImmutable ();
|
|
stateMap_->setImmutable ();
|
|
|
|
if (! setup(config))
|
|
loaded = false;
|
|
|
|
if (! loaded)
|
|
{
|
|
info_.hash = calculateLedgerHash(info_);
|
|
family.missing_node (info_.hash);
|
|
}
|
|
}
|
|
|
|
// Create a new ledger that follows this one
|
|
Ledger::Ledger (Ledger const& prevLedger,
|
|
NetClock::time_point closeTime)
|
|
: mImmutable (false)
|
|
, txMap_ (std::make_shared <SHAMap> (
|
|
SHAMapType::TRANSACTION,
|
|
prevLedger.stateMap_->family(),
|
|
prevLedger.stateMap_->get_version()))
|
|
, stateMap_ (prevLedger.stateMap_->snapShot (true))
|
|
, fees_(prevLedger.fees_)
|
|
, rules_(prevLedger.rules_)
|
|
{
|
|
info_.seq = prevLedger.info_.seq + 1;
|
|
info_.parentCloseTime =
|
|
prevLedger.info_.closeTime;
|
|
info_.hash = prevLedger.info().hash + uint256(1);
|
|
info_.drops = prevLedger.info().drops;
|
|
info_.closeTimeResolution = prevLedger.info_.closeTimeResolution;
|
|
info_.parentHash = prevLedger.info().hash;
|
|
info_.closeTimeResolution = getNextLedgerTimeResolution(
|
|
prevLedger.info_.closeTimeResolution,
|
|
getCloseAgree(prevLedger.info()), info_.seq);
|
|
|
|
if (stateMap_->is_v2())
|
|
{
|
|
info_.closeFlags |= sLCF_SHAMapV2;
|
|
}
|
|
|
|
if (prevLedger.info_.closeTime == NetClock::time_point{})
|
|
{
|
|
info_.closeTime = roundCloseTime(closeTime, info_.closeTimeResolution);
|
|
}
|
|
else
|
|
{
|
|
info_.closeTime =
|
|
prevLedger.info_.closeTime + info_.closeTimeResolution;
|
|
}
|
|
}
|
|
|
|
Ledger::Ledger (
|
|
LedgerInfo const& info,
|
|
Config const& config,
|
|
Family& family)
|
|
: mImmutable (true)
|
|
, txMap_ (std::make_shared <SHAMap> (SHAMapType::TRANSACTION,
|
|
info.txHash, family, SHAMap::version{1}))
|
|
, stateMap_ (std::make_shared <SHAMap> (SHAMapType::STATE,
|
|
info.accountHash, family, SHAMap::version{1}))
|
|
, rules_{config.features}
|
|
, info_ (info)
|
|
{
|
|
info_.hash = calculateLedgerHash (info_);
|
|
}
|
|
|
|
Ledger::Ledger (std::uint32_t ledgerSeq,
|
|
NetClock::time_point closeTime, Config const& config,
|
|
Family& family)
|
|
: mImmutable (false)
|
|
, txMap_ (std::make_shared <SHAMap> (
|
|
SHAMapType::TRANSACTION, family, SHAMap::version{1}))
|
|
, stateMap_ (std::make_shared <SHAMap> (
|
|
SHAMapType::STATE, family, SHAMap::version{1}))
|
|
, rules_{config.features}
|
|
{
|
|
info_.seq = ledgerSeq;
|
|
info_.closeTime = closeTime;
|
|
info_.closeTimeResolution = ledgerDefaultTimeResolution;
|
|
setup(config);
|
|
}
|
|
|
|
void Ledger::setImmutable (Config const& config)
|
|
{
|
|
// Force update, since this is the only
|
|
// place the hash transitions to valid
|
|
if (! mImmutable)
|
|
{
|
|
info_.txHash = txMap_->getHash ().as_uint256();
|
|
info_.accountHash = stateMap_->getHash ().as_uint256();
|
|
}
|
|
|
|
info_.hash = calculateLedgerHash (info_);
|
|
|
|
mImmutable = true;
|
|
txMap_->setImmutable ();
|
|
stateMap_->setImmutable ();
|
|
setup(config);
|
|
}
|
|
|
|
void
|
|
Ledger::setAccepted(NetClock::time_point closeTime,
|
|
NetClock::duration closeResolution,
|
|
bool correctCloseTime, Config const& config)
|
|
{
|
|
// Used when we witnessed the consensus.
|
|
assert (!open());
|
|
|
|
info_.closeTime = closeTime;
|
|
info_.closeTimeResolution = closeResolution;
|
|
info_.closeFlags = correctCloseTime ? 0 : sLCF_NoConsensusTime;
|
|
setImmutable (config);
|
|
}
|
|
|
|
bool Ledger::addSLE (SLE const& sle)
|
|
{
|
|
SHAMapItem item (sle.key(), sle.getSerializer());
|
|
return stateMap_->addItem(std::move(item), false, false);
|
|
}
|
|
|
|
//------------------------------------------------------------------------------
|
|
|
|
std::shared_ptr<STTx const>
|
|
deserializeTx (SHAMapItem const& item)
|
|
{
|
|
SerialIter sit(item.slice());
|
|
return std::make_shared<STTx const>(sit);
|
|
}
|
|
|
|
std::pair<std::shared_ptr<
|
|
STTx const>, std::shared_ptr<
|
|
STObject const>>
|
|
deserializeTxPlusMeta (SHAMapItem const& item)
|
|
{
|
|
std::pair<std::shared_ptr<
|
|
STTx const>, std::shared_ptr<
|
|
STObject const>> result;
|
|
SerialIter sit(item.slice());
|
|
{
|
|
SerialIter s(sit.getSlice(
|
|
sit.getVLDataLength()));
|
|
result.first = std::make_shared<
|
|
STTx const>(s);
|
|
}
|
|
{
|
|
SerialIter s(sit.getSlice(
|
|
sit.getVLDataLength()));
|
|
result.second = std::make_shared<
|
|
STObject const>(s, sfMetadata);
|
|
}
|
|
return result;
|
|
}
|
|
|
|
//------------------------------------------------------------------------------
|
|
|
|
bool
|
|
Ledger::exists (Keylet const& k) const
|
|
{
|
|
// VFALCO NOTE Perhaps check the type for debug builds?
|
|
return stateMap_->hasItem(k.key);
|
|
}
|
|
|
|
boost::optional<uint256>
|
|
Ledger::succ (uint256 const& key,
|
|
boost::optional<uint256> const& last) const
|
|
{
|
|
auto item = stateMap_->upper_bound(key);
|
|
if (item == stateMap_->end())
|
|
return boost::none;
|
|
if (last && item->key() >= last)
|
|
return boost::none;
|
|
return item->key();
|
|
}
|
|
|
|
std::shared_ptr<SLE const>
|
|
Ledger::read (Keylet const& k) const
|
|
{
|
|
if (k.key == zero)
|
|
{
|
|
assert(false);
|
|
return nullptr;
|
|
}
|
|
auto const& item =
|
|
stateMap_->peekItem(k.key);
|
|
if (! item)
|
|
return nullptr;
|
|
auto sle = std::make_shared<SLE>(
|
|
SerialIter{item->data(),
|
|
item->size()}, item->key());
|
|
if (! k.check(*sle))
|
|
return nullptr;
|
|
// need move otherwise makes a copy
|
|
// because return type is different
|
|
return std::move(sle);
|
|
}
|
|
|
|
//------------------------------------------------------------------------------
|
|
|
|
auto
|
|
Ledger::slesBegin() const ->
|
|
std::unique_ptr<sles_type::iter_base>
|
|
{
|
|
return std::make_unique<
|
|
sles_iter_impl>(
|
|
stateMap_->begin(), *this);
|
|
}
|
|
|
|
auto
|
|
Ledger::slesEnd() const ->
|
|
std::unique_ptr<sles_type::iter_base>
|
|
{
|
|
return std::make_unique<
|
|
sles_iter_impl>(
|
|
stateMap_->end(), *this);
|
|
}
|
|
|
|
auto
|
|
Ledger::slesUpperBound(uint256 const& key) const ->
|
|
std::unique_ptr<sles_type::iter_base>
|
|
{
|
|
return std::make_unique<
|
|
sles_iter_impl>(
|
|
stateMap_->upper_bound(key), *this);
|
|
}
|
|
|
|
auto
|
|
Ledger::txsBegin() const ->
|
|
std::unique_ptr<txs_type::iter_base>
|
|
{
|
|
return std::make_unique<txs_iter_impl>(
|
|
!open(), txMap_->begin(), *this);
|
|
}
|
|
|
|
auto
|
|
Ledger::txsEnd() const ->
|
|
std::unique_ptr<txs_type::iter_base>
|
|
{
|
|
return std::make_unique<txs_iter_impl>(
|
|
!open(), txMap_->end(), *this);
|
|
}
|
|
|
|
bool
|
|
Ledger::txExists (uint256 const& key) const
|
|
{
|
|
return txMap_->hasItem (key);
|
|
}
|
|
|
|
auto
|
|
Ledger::txRead(
|
|
key_type const& key) const ->
|
|
tx_type
|
|
{
|
|
auto const& item =
|
|
txMap_->peekItem(key);
|
|
if (! item)
|
|
return {};
|
|
if (!open())
|
|
{
|
|
auto result =
|
|
deserializeTxPlusMeta(*item);
|
|
return { std::move(result.first),
|
|
std::move(result.second) };
|
|
}
|
|
return { deserializeTx(*item), nullptr };
|
|
}
|
|
|
|
auto
|
|
Ledger::digest (key_type const& key) const ->
|
|
boost::optional<digest_type>
|
|
{
|
|
SHAMapHash digest;
|
|
// VFALCO Unfortunately this loads the item
|
|
// from the NodeStore needlessly.
|
|
if (! stateMap_->peekItem(key, digest))
|
|
return boost::none;
|
|
return digest.as_uint256();
|
|
}
|
|
|
|
//------------------------------------------------------------------------------
|
|
|
|
void
|
|
Ledger::rawErase(std::shared_ptr<SLE> const& sle)
|
|
{
|
|
if (! stateMap_->delItem(sle->key()))
|
|
LogicError("Ledger::rawErase: key not found");
|
|
}
|
|
|
|
void
|
|
Ledger::rawInsert(std::shared_ptr<SLE> const& sle)
|
|
{
|
|
Serializer ss;
|
|
sle->add(ss);
|
|
auto item = std::make_shared<
|
|
SHAMapItem const>(sle->key(),
|
|
std::move(ss));
|
|
// VFALCO NOTE addGiveItem should take ownership
|
|
if (! stateMap_->addGiveItem(
|
|
std::move(item), false, false))
|
|
LogicError("Ledger::rawInsert: key already exists");
|
|
}
|
|
|
|
void
|
|
Ledger::rawReplace(std::shared_ptr<SLE> const& sle)
|
|
{
|
|
Serializer ss;
|
|
sle->add(ss);
|
|
auto item = std::make_shared<
|
|
SHAMapItem const>(sle->key(),
|
|
std::move(ss));
|
|
// VFALCO NOTE updateGiveItem should take ownership
|
|
if (! stateMap_->updateGiveItem(
|
|
std::move(item), false, false))
|
|
LogicError("Ledger::rawReplace: key not found");
|
|
}
|
|
|
|
void
|
|
Ledger::rawTxInsert (uint256 const& key,
|
|
std::shared_ptr<Serializer const
|
|
> const& txn, std::shared_ptr<
|
|
Serializer const> const& metaData)
|
|
{
|
|
assert (metaData);
|
|
|
|
// low-level - just add to table
|
|
Serializer s(txn->getDataLength () +
|
|
metaData->getDataLength () + 16);
|
|
s.addVL (txn->peekData ());
|
|
s.addVL (metaData->peekData ());
|
|
auto item = std::make_shared<
|
|
SHAMapItem const> (key, std::move(s));
|
|
if (! txMap().addGiveItem
|
|
(std::move(item), true, true))
|
|
LogicError("duplicate_tx: " + to_string(key));
|
|
}
|
|
|
|
bool
|
|
Ledger::setup (Config const& config)
|
|
{
|
|
bool ret = true;
|
|
|
|
fees_.base = config.FEE_DEFAULT;
|
|
fees_.units = config.TRANSACTION_FEE_BASE;
|
|
fees_.reserve = config.FEE_ACCOUNT_RESERVE;
|
|
fees_.increment = config.FEE_OWNER_RESERVE;
|
|
|
|
try
|
|
{
|
|
if (auto const sle = read(keylet::fees()))
|
|
{
|
|
// VFALCO NOTE Why getFieldIndex and not isFieldPresent?
|
|
|
|
if (sle->getFieldIndex (sfBaseFee) != -1)
|
|
fees_.base = sle->getFieldU64 (sfBaseFee);
|
|
|
|
if (sle->getFieldIndex (sfReferenceFeeUnits) != -1)
|
|
fees_.units = sle->getFieldU32 (sfReferenceFeeUnits);
|
|
|
|
if (sle->getFieldIndex (sfReserveBase) != -1)
|
|
fees_.reserve = sle->getFieldU32 (sfReserveBase);
|
|
|
|
if (sle->getFieldIndex (sfReserveIncrement) != -1)
|
|
fees_.increment = sle->getFieldU32 (sfReserveIncrement);
|
|
}
|
|
}
|
|
catch (SHAMapMissingNode &)
|
|
{
|
|
ret = false;
|
|
}
|
|
catch (std::exception const&)
|
|
{
|
|
Rethrow();
|
|
}
|
|
|
|
try
|
|
{
|
|
rules_ = Rules(*this, config.features);
|
|
}
|
|
catch (SHAMapMissingNode &)
|
|
{
|
|
ret = false;
|
|
}
|
|
catch (std::exception const&)
|
|
{
|
|
Rethrow();
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
std::shared_ptr<SLE>
|
|
Ledger::peek (Keylet const& k) const
|
|
{
|
|
auto const& value =
|
|
stateMap_->peekItem(k.key);
|
|
if (! value)
|
|
return nullptr;
|
|
auto sle = std::make_shared<SLE>(
|
|
SerialIter{value->data(), value->size()}, value->key());
|
|
if (! k.check(*sle))
|
|
return nullptr;
|
|
return sle;
|
|
}
|
|
|
|
//------------------------------------------------------------------------------
|
|
bool Ledger::walkLedger (beast::Journal j) const
|
|
{
|
|
std::vector <SHAMapMissingNode> missingNodes1;
|
|
std::vector <SHAMapMissingNode> missingNodes2;
|
|
|
|
if (stateMap_->getHash().isZero() &&
|
|
! info_.accountHash.isZero() &&
|
|
! stateMap_->fetchRoot (SHAMapHash{info_.accountHash}, nullptr))
|
|
{
|
|
missingNodes1.emplace_back (SHAMapType::STATE, SHAMapHash{info_.accountHash});
|
|
}
|
|
else
|
|
{
|
|
stateMap_->walkMap (missingNodes1, 32);
|
|
}
|
|
|
|
if (!missingNodes1.empty ())
|
|
{
|
|
if (auto stream = j.info())
|
|
{
|
|
stream << missingNodes1.size () << " missing account node(s)";
|
|
stream << "First: " << missingNodes1[0];
|
|
}
|
|
}
|
|
|
|
if (txMap_->getHash().isZero() &&
|
|
info_.txHash.isNonZero() &&
|
|
! txMap_->fetchRoot (SHAMapHash{info_.txHash}, nullptr))
|
|
{
|
|
missingNodes2.emplace_back (SHAMapType::TRANSACTION, SHAMapHash{info_.txHash});
|
|
}
|
|
else
|
|
{
|
|
txMap_->walkMap (missingNodes2, 32);
|
|
}
|
|
|
|
if (!missingNodes2.empty ())
|
|
{
|
|
if (auto stream = j.info())
|
|
{
|
|
stream << missingNodes2.size () << " missing transaction node(s)";
|
|
stream << "First: " << missingNodes2[0];
|
|
}
|
|
}
|
|
return missingNodes1.empty () && missingNodes2.empty ();
|
|
}
|
|
|
|
bool Ledger::assertSane (beast::Journal ledgerJ)
|
|
{
|
|
if (info_.hash.isNonZero () &&
|
|
info_.accountHash.isNonZero () &&
|
|
stateMap_ &&
|
|
txMap_ &&
|
|
(info_.accountHash == stateMap_->getHash ().as_uint256()) &&
|
|
(info_.txHash == txMap_->getHash ().as_uint256()))
|
|
{
|
|
return true;
|
|
}
|
|
|
|
Json::Value j = getJson (*this);
|
|
|
|
j [jss::accountTreeHash] = to_string (info_.accountHash);
|
|
j [jss::transTreeHash] = to_string (info_.txHash);
|
|
|
|
JLOG (ledgerJ.fatal()) << "ledger is not sane" << j;
|
|
|
|
assert (false);
|
|
|
|
return false;
|
|
}
|
|
|
|
// update the skip list with the information from our previous ledger
|
|
// VFALCO TODO Document this skip list concept
|
|
void Ledger::updateSkipList ()
|
|
{
|
|
if (info_.seq == 0) // genesis ledger has no previous ledger
|
|
return;
|
|
|
|
std::uint32_t prevIndex = info_.seq - 1;
|
|
|
|
// update record of every 256th ledger
|
|
if ((prevIndex & 0xff) == 0)
|
|
{
|
|
auto const k = keylet::skip(prevIndex);
|
|
auto sle = peek(k);
|
|
std::vector<uint256> hashes;
|
|
|
|
bool created;
|
|
if (! sle)
|
|
{
|
|
sle = std::make_shared<SLE>(k);
|
|
created = true;
|
|
}
|
|
else
|
|
{
|
|
hashes = static_cast<decltype(hashes)>(
|
|
sle->getFieldV256(sfHashes));
|
|
created = false;
|
|
}
|
|
|
|
assert (hashes.size () <= 256);
|
|
hashes.push_back (info_.parentHash);
|
|
sle->setFieldV256 (sfHashes, STVector256 (hashes));
|
|
sle->setFieldU32 (sfLastLedgerSequence, prevIndex);
|
|
if (created)
|
|
rawInsert(sle);
|
|
else
|
|
rawReplace(sle);
|
|
}
|
|
|
|
// update record of past 256 ledger
|
|
auto const k = keylet::skip();
|
|
auto sle = peek(k);
|
|
std::vector <uint256> hashes;
|
|
bool created;
|
|
if (! sle)
|
|
{
|
|
sle = std::make_shared<SLE>(k);
|
|
created = true;
|
|
}
|
|
else
|
|
{
|
|
hashes = static_cast<decltype(hashes)>(
|
|
sle->getFieldV256 (sfHashes));
|
|
created = false;
|
|
}
|
|
assert (hashes.size () <= 256);
|
|
if (hashes.size () == 256)
|
|
hashes.erase (hashes.begin ());
|
|
hashes.push_back (info_.parentHash);
|
|
sle->setFieldV256 (sfHashes, STVector256 (hashes));
|
|
sle->setFieldU32 (sfLastLedgerSequence, prevIndex);
|
|
if (created)
|
|
rawInsert(sle);
|
|
else
|
|
rawReplace(sle);
|
|
}
|
|
|
|
static bool saveValidatedLedger (
|
|
Application& app,
|
|
std::shared_ptr<Ledger const> const& ledger,
|
|
bool current)
|
|
{
|
|
auto j = app.journal ("Ledger");
|
|
|
|
if (! app.pendingSaves().startWork (ledger->info().seq))
|
|
{
|
|
// The save was completed synchronously
|
|
JLOG (j.debug()) << "Save aborted";
|
|
return true;
|
|
}
|
|
|
|
// TODO(tom): Fix this hard-coded SQL!
|
|
JLOG (j.trace())
|
|
<< "saveValidatedLedger "
|
|
<< (current ? "" : "fromAcquire ") << ledger->info().seq;
|
|
static boost::format deleteLedger (
|
|
"DELETE FROM Ledgers WHERE LedgerSeq = %u;");
|
|
static boost::format deleteTrans1 (
|
|
"DELETE FROM Transactions WHERE LedgerSeq = %u;");
|
|
static boost::format deleteTrans2 (
|
|
"DELETE FROM AccountTransactions WHERE LedgerSeq = %u;");
|
|
static boost::format deleteAcctTrans (
|
|
"DELETE FROM AccountTransactions WHERE TransID = '%s';");
|
|
|
|
auto seq = ledger->info().seq;
|
|
|
|
if (! ledger->info().accountHash.isNonZero ())
|
|
{
|
|
JLOG (j.fatal()) << "AH is zero: "
|
|
<< getJson (*ledger);
|
|
assert (false);
|
|
}
|
|
|
|
if (ledger->info().accountHash != ledger->stateMap().getHash ().as_uint256())
|
|
{
|
|
JLOG (j.fatal()) << "sAL: " << ledger->info().accountHash
|
|
<< " != " << ledger->stateMap().getHash ();
|
|
JLOG (j.fatal()) << "saveAcceptedLedger: seq="
|
|
<< seq << ", current=" << current;
|
|
assert (false);
|
|
}
|
|
|
|
assert (ledger->info().txHash == ledger->txMap().getHash ().as_uint256());
|
|
|
|
// Save the ledger header in the hashed object store
|
|
{
|
|
Serializer s (128);
|
|
s.add32 (HashPrefix::ledgerMaster);
|
|
addRaw(ledger->info(), s);
|
|
app.getNodeStore ().store (
|
|
hotLEDGER, std::move (s.modData ()), ledger->info().hash);
|
|
}
|
|
|
|
|
|
AcceptedLedger::pointer aLedger;
|
|
try
|
|
{
|
|
aLedger = app.getAcceptedLedgerCache().fetch (ledger->info().hash);
|
|
if (! aLedger)
|
|
{
|
|
aLedger = std::make_shared<AcceptedLedger>(ledger, app.accountIDCache(), app.logs());
|
|
app.getAcceptedLedgerCache().canonicalize(ledger->info().hash, aLedger);
|
|
}
|
|
}
|
|
catch (std::exception const&)
|
|
{
|
|
JLOG (j.warn()) << "An accepted ledger was missing nodes";
|
|
app.getLedgerMaster().failedSave(seq, ledger->info().hash);
|
|
// Clients can now trust the database for information about this
|
|
// ledger sequence.
|
|
app.pendingSaves().finishWork(seq);
|
|
return false;
|
|
}
|
|
|
|
{
|
|
auto db = app.getLedgerDB ().checkoutDb();
|
|
*db << boost::str (deleteLedger % seq);
|
|
}
|
|
|
|
{
|
|
auto db = app.getTxnDB ().checkoutDb ();
|
|
|
|
soci::transaction tr(*db);
|
|
|
|
*db << boost::str (deleteTrans1 % seq);
|
|
*db << boost::str (deleteTrans2 % seq);
|
|
|
|
std::string const ledgerSeq (std::to_string (seq));
|
|
|
|
for (auto const& vt : aLedger->getMap ())
|
|
{
|
|
uint256 transactionID = vt.second->getTransactionID ();
|
|
|
|
app.getMasterTransaction ().inLedger (
|
|
transactionID, seq);
|
|
|
|
std::string const txnId (to_string (transactionID));
|
|
std::string const txnSeq (std::to_string (vt.second->getTxnSeq ()));
|
|
|
|
*db << boost::str (deleteAcctTrans % transactionID);
|
|
|
|
auto const& accts = vt.second->getAffected ();
|
|
|
|
if (!accts.empty ())
|
|
{
|
|
std::string sql (
|
|
"INSERT INTO AccountTransactions "
|
|
"(TransID, Account, LedgerSeq, TxnSeq) VALUES ");
|
|
|
|
// Try to make an educated guess on how much space we'll need
|
|
// for our arguments. In argument order we have:
|
|
// 64 + 34 + 10 + 10 = 118 + 10 extra = 128 bytes
|
|
sql.reserve (sql.length () + (accts.size () * 128));
|
|
|
|
bool first = true;
|
|
for (auto const& account : accts)
|
|
{
|
|
if (!first)
|
|
sql += ", ('";
|
|
else
|
|
{
|
|
sql += "('";
|
|
first = false;
|
|
}
|
|
|
|
sql += txnId;
|
|
sql += "','";
|
|
sql += app.accountIDCache().toBase58(account);
|
|
sql += "',";
|
|
sql += ledgerSeq;
|
|
sql += ",";
|
|
sql += txnSeq;
|
|
sql += ")";
|
|
}
|
|
sql += ";";
|
|
JLOG (j.trace()) << "ActTx: " << sql;
|
|
*db << sql;
|
|
}
|
|
else
|
|
{
|
|
JLOG (j.warn())
|
|
<< "Transaction in ledger " << seq
|
|
<< " affects no accounts";
|
|
JLOG (j.warn())
|
|
<< vt.second->getTxn()->getJson(0);
|
|
}
|
|
|
|
*db <<
|
|
(STTx::getMetaSQLInsertReplaceHeader () +
|
|
vt.second->getTxn ()->getMetaSQL (
|
|
seq, vt.second->getEscMeta ()) + ";");
|
|
}
|
|
|
|
tr.commit ();
|
|
}
|
|
|
|
{
|
|
static std::string addLedger(
|
|
R"sql(INSERT OR REPLACE INTO Ledgers
|
|
(LedgerHash,LedgerSeq,PrevHash,TotalCoins,ClosingTime,PrevClosingTime,
|
|
CloseTimeRes,CloseFlags,AccountSetHash,TransSetHash)
|
|
VALUES
|
|
(:ledgerHash,:ledgerSeq,:prevHash,:totalCoins,:closingTime,:prevClosingTime,
|
|
:closeTimeRes,:closeFlags,:accountSetHash,:transSetHash);)sql");
|
|
static std::string updateVal(
|
|
R"sql(UPDATE Validations SET LedgerSeq = :ledgerSeq, InitialSeq = :initialSeq
|
|
WHERE LedgerHash = :ledgerHash;)sql");
|
|
|
|
auto db (app.getLedgerDB ().checkoutDb ());
|
|
|
|
soci::transaction tr(*db);
|
|
|
|
auto const hash = to_string (ledger->info().hash);
|
|
auto const parentHash = to_string (ledger->info().parentHash);
|
|
auto const drops = to_string (ledger->info().drops);
|
|
auto const closeTime =
|
|
ledger->info().closeTime.time_since_epoch().count();
|
|
auto const parentCloseTime =
|
|
ledger->info().parentCloseTime.time_since_epoch().count();
|
|
auto const closeTimeResolution =
|
|
ledger->info().closeTimeResolution.count();
|
|
auto const closeFlags = ledger->info().closeFlags;
|
|
auto const accountHash = to_string (ledger->info().accountHash);
|
|
auto const txHash = to_string (ledger->info().txHash);
|
|
|
|
*db << addLedger,
|
|
soci::use(hash),
|
|
soci::use(seq),
|
|
soci::use(parentHash),
|
|
soci::use(drops),
|
|
soci::use(closeTime),
|
|
soci::use(parentCloseTime),
|
|
soci::use(closeTimeResolution),
|
|
soci::use(closeFlags),
|
|
soci::use(accountHash),
|
|
soci::use(txHash);
|
|
|
|
|
|
*db << updateVal,
|
|
soci::use(seq),
|
|
soci::use(seq),
|
|
soci::use(hash);
|
|
|
|
tr.commit();
|
|
}
|
|
|
|
// Clients can now trust the database for
|
|
// information about this ledger sequence.
|
|
app.pendingSaves().finishWork(seq);
|
|
return true;
|
|
}
|
|
|
|
/** Save, or arrange to save, a fully-validated ledger
|
|
Returns false on error
|
|
*/
|
|
bool pendSaveValidated (
|
|
Application& app,
|
|
std::shared_ptr<Ledger const> const& ledger,
|
|
bool isSynchronous,
|
|
bool isCurrent)
|
|
{
|
|
if (! app.getHashRouter ().setFlags (ledger->info().hash, SF_SAVED))
|
|
{
|
|
// We have tried to save this ledger recently
|
|
auto stream = app.journal ("Ledger").debug();
|
|
JLOG (stream) << "Double pend save for "
|
|
<< ledger->info().seq;
|
|
|
|
if (! isSynchronous ||
|
|
! app.pendingSaves().pending (ledger->info().seq))
|
|
{
|
|
// Either we don't need it to be finished
|
|
// or it is finished
|
|
return true;
|
|
}
|
|
}
|
|
|
|
assert (ledger->isImmutable ());
|
|
|
|
if (! app.pendingSaves().shouldWork (ledger->info().seq, isSynchronous))
|
|
{
|
|
auto stream = app.journal ("Ledger").debug();
|
|
JLOG (stream)
|
|
<< "Pend save with seq in pending saves "
|
|
<< ledger->info().seq;
|
|
|
|
return true;
|
|
}
|
|
|
|
if (isSynchronous)
|
|
return saveValidatedLedger(app, ledger, isCurrent);
|
|
|
|
auto job = [ledger, &app, isCurrent] (Job&) {
|
|
saveValidatedLedger(app, ledger, isCurrent);
|
|
};
|
|
|
|
if (isCurrent)
|
|
{
|
|
app.getJobQueue().addJob(
|
|
jtPUBLEDGER, "Ledger::pendSave", job);
|
|
}
|
|
else
|
|
{
|
|
app.getJobQueue ().addJob(
|
|
jtPUBOLDLEDGER, "Ledger::pendOldSave", job);
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
void
|
|
Ledger::make_v2()
|
|
{
|
|
assert (! mImmutable);
|
|
stateMap_ = stateMap_->make_v2();
|
|
txMap_ = txMap_->make_v2();
|
|
info_.validated = false;
|
|
info_.accountHash = stateMap_->getHash ().as_uint256();
|
|
info_.txHash = txMap_->getHash ().as_uint256();
|
|
info_.hash = calculateLedgerHash (info_);
|
|
info_.closeFlags |= sLCF_SHAMapV2;
|
|
}
|
|
|
|
void
|
|
Ledger::unshare() const
|
|
{
|
|
stateMap_->unshare();
|
|
txMap_->unshare();
|
|
}
|
|
|
|
void
|
|
Ledger::invariants() const
|
|
{
|
|
stateMap_->invariants();
|
|
txMap_->invariants();
|
|
}
|
|
|
|
//------------------------------------------------------------------------------
|
|
|
|
/*
|
|
* Load a ledger from the database.
|
|
*
|
|
* @param sqlSuffix: Additional string to append to the sql query.
|
|
* (typically a where clause).
|
|
* @return The ledger, ledger sequence, and ledger hash.
|
|
*/
|
|
std::tuple<std::shared_ptr<Ledger>, std::uint32_t, uint256>
|
|
loadLedgerHelper(std::string const& sqlSuffix, Application& app)
|
|
{
|
|
uint256 ledgerHash{};
|
|
std::uint32_t ledgerSeq{0};
|
|
|
|
auto db = app.getLedgerDB ().checkoutDb ();
|
|
|
|
boost::optional<std::string> sLedgerHash, sPrevHash, sAccountHash,
|
|
sTransHash;
|
|
boost::optional<std::uint64_t> totDrops, closingTime, prevClosingTime,
|
|
closeResolution, closeFlags, ledgerSeq64;
|
|
|
|
std::string const sql =
|
|
"SELECT "
|
|
"LedgerHash, PrevHash, AccountSetHash, TransSetHash, "
|
|
"TotalCoins,"
|
|
"ClosingTime, PrevClosingTime, CloseTimeRes, CloseFlags,"
|
|
"LedgerSeq from Ledgers " +
|
|
sqlSuffix + ";";
|
|
|
|
*db << sql,
|
|
soci::into(sLedgerHash),
|
|
soci::into(sPrevHash),
|
|
soci::into(sAccountHash),
|
|
soci::into(sTransHash),
|
|
soci::into(totDrops),
|
|
soci::into(closingTime),
|
|
soci::into(prevClosingTime),
|
|
soci::into(closeResolution),
|
|
soci::into(closeFlags),
|
|
soci::into(ledgerSeq64);
|
|
|
|
if (!db->got_data ())
|
|
{
|
|
auto stream = app.journal("Ledger").debug();
|
|
JLOG (stream) << "Ledger not found: " << sqlSuffix;
|
|
return std::make_tuple (
|
|
std::shared_ptr<Ledger>(),
|
|
ledgerSeq,
|
|
ledgerHash);
|
|
}
|
|
|
|
ledgerSeq =
|
|
rangeCheckedCast<std::uint32_t>(ledgerSeq64.value_or (0));
|
|
|
|
uint256 prevHash{}, accountHash{}, transHash{};
|
|
if (sLedgerHash)
|
|
ledgerHash.SetHexExact (*sLedgerHash);
|
|
if (sPrevHash)
|
|
prevHash.SetHexExact (*sPrevHash);
|
|
if (sAccountHash)
|
|
accountHash.SetHexExact (*sAccountHash);
|
|
if (sTransHash)
|
|
transHash.SetHexExact (*sTransHash);
|
|
|
|
using time_point = NetClock::time_point;
|
|
using duration = NetClock::duration;
|
|
|
|
LedgerInfo info;
|
|
info.parentHash = prevHash;
|
|
info.txHash = transHash;
|
|
info.accountHash = accountHash;
|
|
info.drops = totDrops.value_or(0);
|
|
info.closeTime = time_point{duration{closingTime.value_or(0)}};
|
|
info.parentCloseTime = time_point{duration{prevClosingTime.value_or(0)}};
|
|
info.closeFlags = closeFlags.value_or(0);
|
|
info.closeTimeResolution = duration{closeResolution.value_or(0)};
|
|
info.seq = ledgerSeq;
|
|
|
|
bool loaded = false;
|
|
|
|
auto ledger = std::make_shared<Ledger>(
|
|
info,
|
|
loaded,
|
|
app.config(),
|
|
app.family(),
|
|
app.journal("Ledger"));
|
|
|
|
if (!loaded)
|
|
ledger.reset();
|
|
|
|
return std::make_tuple (ledger, ledgerSeq, ledgerHash);
|
|
}
|
|
|
|
static
|
|
void finishLoadByIndexOrHash(
|
|
std::shared_ptr<Ledger> const& ledger,
|
|
Config const& config,
|
|
beast::Journal j)
|
|
{
|
|
if (!ledger)
|
|
return;
|
|
|
|
ledger->setImmutable (config);
|
|
|
|
JLOG (j.trace())
|
|
<< "Loaded ledger: " << to_string (ledger->info().hash);
|
|
|
|
ledger->setFull ();
|
|
}
|
|
|
|
std::shared_ptr<Ledger>
|
|
loadByIndex (std::uint32_t ledgerIndex, Application& app)
|
|
{
|
|
std::shared_ptr<Ledger> ledger;
|
|
{
|
|
std::ostringstream s;
|
|
s << "WHERE LedgerSeq = " << ledgerIndex;
|
|
std::tie (ledger, std::ignore, std::ignore) =
|
|
loadLedgerHelper (s.str (), app);
|
|
}
|
|
|
|
finishLoadByIndexOrHash (ledger, app.config(),
|
|
app.journal ("Ledger"));
|
|
return ledger;
|
|
}
|
|
|
|
std::shared_ptr<Ledger>
|
|
loadByHash (uint256 const& ledgerHash, Application& app)
|
|
{
|
|
std::shared_ptr<Ledger> ledger;
|
|
{
|
|
std::ostringstream s;
|
|
s << "WHERE LedgerHash = '" << ledgerHash << "'";
|
|
std::tie (ledger, std::ignore, std::ignore) =
|
|
loadLedgerHelper (s.str (), app);
|
|
}
|
|
|
|
finishLoadByIndexOrHash (ledger, app.config(),
|
|
app.journal ("Ledger"));
|
|
|
|
assert (!ledger || ledger->info().hash == ledgerHash);
|
|
|
|
return ledger;
|
|
}
|
|
|
|
uint256
|
|
getHashByIndex (std::uint32_t ledgerIndex, Application& app)
|
|
{
|
|
uint256 ret;
|
|
|
|
std::string sql =
|
|
"SELECT LedgerHash FROM Ledgers INDEXED BY SeqLedger WHERE LedgerSeq='";
|
|
sql.append (beast::lexicalCastThrow <std::string> (ledgerIndex));
|
|
sql.append ("';");
|
|
|
|
std::string hash;
|
|
{
|
|
auto db = app.getLedgerDB ().checkoutDb ();
|
|
|
|
boost::optional<std::string> lh;
|
|
*db << sql,
|
|
soci::into (lh);
|
|
|
|
if (!db->got_data () || !lh)
|
|
return ret;
|
|
|
|
hash = *lh;
|
|
if (hash.empty ())
|
|
return ret;
|
|
}
|
|
|
|
ret.SetHexExact (hash);
|
|
return ret;
|
|
}
|
|
|
|
bool
|
|
getHashesByIndex(std::uint32_t ledgerIndex,
|
|
uint256& ledgerHash, uint256& parentHash,
|
|
Application& app)
|
|
{
|
|
auto db = app.getLedgerDB ().checkoutDb ();
|
|
|
|
boost::optional <std::string> lhO, phO;
|
|
|
|
*db << "SELECT LedgerHash,PrevHash FROM Ledgers "
|
|
"INDEXED BY SeqLedger Where LedgerSeq = :ls;",
|
|
soci::into (lhO),
|
|
soci::into (phO),
|
|
soci::use (ledgerIndex);
|
|
|
|
if (!lhO || !phO)
|
|
{
|
|
auto stream = app.journal ("Ledger").trace();
|
|
JLOG (stream)
|
|
<< "Don't have ledger " << ledgerIndex;
|
|
return false;
|
|
}
|
|
|
|
ledgerHash.SetHexExact (*lhO);
|
|
parentHash.SetHexExact (*phO);
|
|
|
|
return true;
|
|
}
|
|
|
|
std::map< std::uint32_t, std::pair<uint256, uint256> >
|
|
getHashesByIndex (std::uint32_t minSeq, std::uint32_t maxSeq,
|
|
Application& app)
|
|
{
|
|
std::map< std::uint32_t, std::pair<uint256, uint256> > ret;
|
|
|
|
std::string sql =
|
|
"SELECT LedgerSeq,LedgerHash,PrevHash FROM Ledgers WHERE LedgerSeq >= ";
|
|
sql.append (beast::lexicalCastThrow <std::string> (minSeq));
|
|
sql.append (" AND LedgerSeq <= ");
|
|
sql.append (beast::lexicalCastThrow <std::string> (maxSeq));
|
|
sql.append (";");
|
|
|
|
auto db = app.getLedgerDB ().checkoutDb ();
|
|
|
|
std::uint64_t ls;
|
|
std::string lh;
|
|
boost::optional<std::string> ph;
|
|
soci::statement st =
|
|
(db->prepare << sql,
|
|
soci::into (ls),
|
|
soci::into (lh),
|
|
soci::into (ph));
|
|
|
|
st.execute ();
|
|
while (st.fetch ())
|
|
{
|
|
std::pair<uint256, uint256>& hashes =
|
|
ret[rangeCheckedCast<std::uint32_t>(ls)];
|
|
hashes.first.SetHexExact (lh);
|
|
if (ph)
|
|
hashes.second.SetHexExact (*ph);
|
|
else
|
|
hashes.second.zero ();
|
|
if (!ph)
|
|
{
|
|
auto stream = app.journal ("Ledger").warn();
|
|
JLOG (stream)
|
|
<< "Null prev hash for ledger seq: " << ls;
|
|
}
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
} // ripple
|