Files
xahaud/src/ripple/app/paths/PathRequests.cpp
Scott Schurr efe3700f70 Don't use JobQueue during shutdown (RIPD-1356):
If the JobQueue is used during shutdown then those Jobs may access
Stoppables after they have already stopped.  This violates the
preconditions of Stoppables and may lead to undefined behavior.

The solution taken here is to reference count all Jobs in the
JobQueue.  At stop time all Jobs already in the JobQueue are
allowed to run to completion, but no further Jobs are allowed
into the JobQueue.

If a Job is rejected from the JobQueue (because we are stopping),
then JobQueue::addJob() returns false, so the caller can make any
necessary adjustments.
2017-07-31 18:39:58 -04:00

280 lines
8.7 KiB
C++

//------------------------------------------------------------------------------
/*
This file is part of rippled: https://github.com/ripple/rippled
Copyright (c) 2012, 2013 Ripple Labs Inc.
Permission to use, copy, modify, and/or distribute this software for any
purpose with or without fee is hereby granted, provided that the above
copyright notice and this permission notice appear in all copies.
THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
ANY SPECIAL , DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
*/
//==============================================================================
#include <BeastConfig.h>
#include <ripple/app/paths/PathRequests.h>
#include <ripple/app/ledger/LedgerMaster.h>
#include <ripple/app/main/Application.h>
#include <ripple/basics/Log.h>
#include <ripple/core/JobQueue.h>
#include <ripple/net/RPCErr.h>
#include <ripple/protocol/ErrorCodes.h>
#include <ripple/protocol/JsonFields.h>
#include <ripple/resource/Fees.h>
#include <algorithm>
namespace ripple {
/** Get the current RippleLineCache, updating it if necessary.
Get the correct ledger to use.
*/
std::shared_ptr<RippleLineCache>
PathRequests::getLineCache (
std::shared_ptr <ReadView const> const& ledger,
bool authoritative)
{
ScopedLockType sl (mLock);
std::uint32_t lineSeq = mLineCache ? mLineCache->getLedger()->seq() : 0;
std::uint32_t lgrSeq = ledger->seq();
if ( (lineSeq == 0) || // no ledger
(authoritative && (lgrSeq > lineSeq)) || // newer authoritative ledger
(authoritative && ((lgrSeq + 8) < lineSeq)) || // we jumped way back for some reason
(lgrSeq > (lineSeq + 8))) // we jumped way forward for some reason
{
mLineCache = std::make_shared<RippleLineCache> (ledger);
}
return mLineCache;
}
void PathRequests::updateAll (std::shared_ptr <ReadView const> const& inLedger,
Job::CancelCallback shouldCancel)
{
auto event =
app_.getJobQueue().makeLoadEvent(
jtPATH_FIND, "PathRequest::updateAll");
std::vector<PathRequest::wptr> requests;
std::shared_ptr<RippleLineCache> cache;
// Get the ledger and cache we should be using
{
ScopedLockType sl (mLock);
requests = requests_;
cache = getLineCache (inLedger, true);
}
bool newRequests = app_.getLedgerMaster().isNewPathRequest();
bool mustBreak = false;
JLOG (mJournal.trace()) <<
"updateAll seq=" << cache->getLedger()->seq() <<
", " << requests.size() << " requests";
int processed = 0, removed = 0;
do
{
for (auto const& wr : requests)
{
if (shouldCancel())
break;
auto request = wr.lock ();
bool remove = true;
if (request)
{
if (!request->needsUpdate (newRequests, cache->getLedger()->seq()))
remove = false;
else
{
if (auto ipSub = request->getSubscriber ())
{
if (!ipSub->getConsumer ().warn ())
{
Json::Value update = request->doUpdate (cache, false);
request->updateComplete ();
update[jss::type] = "path_find";
ipSub->send (update, false);
remove = false;
++processed;
}
}
else if (request->hasCompletion ())
{
// One-shot request with completion function
request->doUpdate (cache, false);
request->updateComplete();
++processed;
}
}
}
if (remove)
{
ScopedLockType sl (mLock);
// Remove any dangling weak pointers or weak
// pointers that refer to this path request.
auto ret = std::remove_if (
requests_.begin(), requests_.end(),
[&removed,&request](auto const& wl)
{
auto r = wl.lock();
if (r && r != request)
return false;
++removed;
return true;
});
requests_.erase (ret, requests_.end());
}
mustBreak = !newRequests &&
app_.getLedgerMaster().isNewPathRequest();
// We weren't handling new requests and then
// there was a new request
if (mustBreak)
break;
}
if (mustBreak)
{ // a new request came in while we were working
newRequests = true;
}
else if (newRequests)
{ // we only did new requests, so we always need a last pass
newRequests = app_.getLedgerMaster().isNewPathRequest();
}
else
{ // if there are no new requests, we are done
newRequests = app_.getLedgerMaster().isNewPathRequest();
if (!newRequests)
break;
}
{
// Get the latest requests, cache, and ledger for next pass
ScopedLockType sl (mLock);
if (requests_.empty())
break;
requests = requests_;
cache = getLineCache (cache->getLedger(), false);
}
}
while (!shouldCancel ());
JLOG (mJournal.debug()) <<
"updateAll complete: " << processed << " processed and " <<
removed << " removed";
}
void PathRequests::insertPathRequest (
PathRequest::pointer const& req)
{
ScopedLockType sl (mLock);
// Insert after any older unserviced requests but before
// any serviced requests
auto ret = std::find_if (
requests_.begin(), requests_.end(),
[](auto const& wl)
{
auto r = wl.lock();
// We come before handled requests
return r && !r->isNew();
});
requests_.emplace (ret, req);
}
// Make a new-style path_find request
Json::Value
PathRequests::makePathRequest(
std::shared_ptr <InfoSub> const& subscriber,
std::shared_ptr<ReadView const> const& inLedger,
Json::Value const& requestJson)
{
auto req = std::make_shared<PathRequest> (
app_, subscriber, ++mLastIdentifier, *this, mJournal);
auto result = req->doCreate (
getLineCache (inLedger, false), requestJson);
if (result.first)
{
subscriber->setPathRequest (req);
insertPathRequest (req);
app_.getLedgerMaster().newPathRequest();
}
return std::move (result.second);
}
// Make an old-style ripple_path_find request
Json::Value
PathRequests::makeLegacyPathRequest(
PathRequest::pointer& req,
std::function <void (void)> completion,
Resource::Consumer& consumer,
std::shared_ptr<ReadView const> const& inLedger,
Json::Value const& request)
{
// This assignment must take place before the
// completion function is called
req = std::make_shared<PathRequest> (
app_, completion, consumer, ++mLastIdentifier,
*this, mJournal);
auto result = req->doCreate (
getLineCache (inLedger, false), request);
if (!result.first)
{
req.reset();
}
else
{
insertPathRequest (req);
if (! app_.getLedgerMaster().newPathRequest())
{
// The newPathRequest failed. Tell the caller.
result.second = rpcError (rpcTOO_BUSY);
req.reset();
}
}
return std::move (result.second);
}
Json::Value
PathRequests::doLegacyPathRequest (
Resource::Consumer& consumer,
std::shared_ptr<ReadView const> const& inLedger,
Json::Value const& request)
{
auto cache = std::make_shared<RippleLineCache> (inLedger);
auto req = std::make_shared<PathRequest> (app_, []{},
consumer, ++mLastIdentifier, *this, mJournal);
auto result = req->doCreate (cache, request);
if (result.first)
result.second = req->doUpdate (cache, false);
return std::move (result.second);
}
} // ripple