mirror of
https://github.com/XRPLF/clio.git
synced 2026-02-04 22:15:32 +00:00
228 lines
7.0 KiB
C++
228 lines
7.0 KiB
C++
//------------------------------------------------------------------------------
|
|
/*
|
|
This file is part of clio: https://github.com/XRPLF/clio
|
|
Copyright (c) 2025, the clio developers.
|
|
|
|
Permission to use, copy, modify, and distribute this software for any
|
|
purpose with or without fee is hereby granted, provided that the above
|
|
copyright notice and this permission notice appear in all copies.
|
|
|
|
THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
|
|
WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
|
|
MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
|
|
ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
|
|
WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
|
|
ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
|
|
OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
|
|
*/
|
|
//==============================================================================
|
|
|
|
#include "cluster/ClusterCommunicationService.hpp"
|
|
|
|
#include "cluster/ClioNode.hpp"
|
|
#include "data/BackendInterface.hpp"
|
|
#include "util/Assert.hpp"
|
|
#include "util/Spawn.hpp"
|
|
#include "util/log/Logger.hpp"
|
|
|
|
#include <boost/asio/bind_cancellation_slot.hpp>
|
|
#include <boost/asio/cancellation_type.hpp>
|
|
#include <boost/asio/error.hpp>
|
|
#include <boost/asio/spawn.hpp>
|
|
#include <boost/asio/steady_timer.hpp>
|
|
#include <boost/asio/use_future.hpp>
|
|
#include <boost/json/parse.hpp>
|
|
#include <boost/json/serialize.hpp>
|
|
#include <boost/json/value.hpp>
|
|
#include <boost/json/value_from.hpp>
|
|
#include <boost/json/value_to.hpp>
|
|
#include <boost/uuid/random_generator.hpp>
|
|
#include <boost/uuid/uuid.hpp>
|
|
|
|
#include <chrono>
|
|
#include <ctime>
|
|
#include <latch>
|
|
#include <memory>
|
|
#include <string>
|
|
#include <utility>
|
|
#include <vector>
|
|
|
|
namespace {
|
|
constexpr auto kTOTAL_WORKERS = 2uz; // 1 reading and 1 writing worker (coroutines)
|
|
} // namespace
|
|
|
|
namespace cluster {
|
|
|
|
ClusterCommunicationService::ClusterCommunicationService(
|
|
std::shared_ptr<data::BackendInterface> backend,
|
|
std::chrono::steady_clock::duration readInterval,
|
|
std::chrono::steady_clock::duration writeInterval
|
|
)
|
|
: backend_(std::move(backend))
|
|
, readInterval_(readInterval)
|
|
, writeInterval_(writeInterval)
|
|
, finishedCountdown_(kTOTAL_WORKERS)
|
|
, selfData_{ClioNode{
|
|
.uuid = std::make_shared<boost::uuids::uuid>(boost::uuids::random_generator{}()),
|
|
.updateTime = std::chrono::system_clock::time_point{}
|
|
}}
|
|
{
|
|
nodesInClusterMetric_.set(1); // The node always sees itself
|
|
isHealthy_ = true;
|
|
}
|
|
|
|
void
|
|
ClusterCommunicationService::run()
|
|
{
|
|
ASSERT(not running_ and not stopped_, "Can only be ran once");
|
|
running_ = true;
|
|
|
|
util::spawn(strand_, [this](boost::asio::yield_context yield) {
|
|
boost::asio::steady_timer timer(yield.get_executor());
|
|
boost::system::error_code ec;
|
|
|
|
while (running_) {
|
|
timer.expires_after(readInterval_);
|
|
auto token = cancelSignal_.slot();
|
|
timer.async_wait(boost::asio::bind_cancellation_slot(token, yield[ec]));
|
|
|
|
if (ec == boost::asio::error::operation_aborted or not running_)
|
|
break;
|
|
|
|
doRead(yield);
|
|
}
|
|
|
|
finishedCountdown_.count_down(1);
|
|
});
|
|
|
|
util::spawn(strand_, [this](boost::asio::yield_context yield) {
|
|
boost::asio::steady_timer timer(yield.get_executor());
|
|
boost::system::error_code ec;
|
|
|
|
while (running_) {
|
|
doWrite();
|
|
timer.expires_after(writeInterval_);
|
|
auto token = cancelSignal_.slot();
|
|
timer.async_wait(boost::asio::bind_cancellation_slot(token, yield[ec]));
|
|
|
|
if (ec == boost::asio::error::operation_aborted or not running_)
|
|
break;
|
|
}
|
|
|
|
finishedCountdown_.count_down(1);
|
|
});
|
|
}
|
|
|
|
ClusterCommunicationService::~ClusterCommunicationService()
|
|
{
|
|
stop();
|
|
}
|
|
|
|
void
|
|
ClusterCommunicationService::stop()
|
|
{
|
|
if (stopped_)
|
|
return;
|
|
|
|
stopped_ = true;
|
|
|
|
// for ASAN to see through concurrency correctly we need to exit all coroutines before joining the ctx
|
|
running_ = false;
|
|
|
|
// cancelSignal_ is not thread safe so we execute emit on the same strand
|
|
boost::asio::spawn(
|
|
strand_, [this](auto&&) { cancelSignal_.emit(boost::asio::cancellation_type::all); }, boost::asio::use_future
|
|
)
|
|
.wait();
|
|
finishedCountdown_.wait();
|
|
|
|
ctx_.join();
|
|
}
|
|
|
|
std::shared_ptr<boost::uuids::uuid>
|
|
ClusterCommunicationService::selfUuid() const
|
|
{
|
|
// Uuid never changes so it is safe to copy it without using strand_
|
|
return selfData_.uuid;
|
|
}
|
|
|
|
ClioNode
|
|
ClusterCommunicationService::selfData() const
|
|
{
|
|
ClioNode result{};
|
|
util::spawn(strand_, [this, &result](boost::asio::yield_context) { result = selfData_; });
|
|
return result;
|
|
}
|
|
|
|
std::expected<std::vector<ClioNode>, std::string>
|
|
ClusterCommunicationService::clusterData() const
|
|
{
|
|
if (not isHealthy_) {
|
|
return std::unexpected{"Service is not healthy"};
|
|
}
|
|
std::vector<ClioNode> result;
|
|
util::spawn(strand_, [this, &result](boost::asio::yield_context) {
|
|
result = otherNodesData_;
|
|
result.push_back(selfData_);
|
|
});
|
|
return result;
|
|
}
|
|
|
|
void
|
|
ClusterCommunicationService::doRead(boost::asio::yield_context yield)
|
|
{
|
|
otherNodesData_.clear();
|
|
|
|
BackendInterface::ClioNodesDataFetchResult expectedResult;
|
|
try {
|
|
expectedResult = backend_->fetchClioNodesData(yield);
|
|
} catch (...) {
|
|
expectedResult = std::unexpected{"Failed to fecth Clio nodes data"};
|
|
}
|
|
|
|
if (!expectedResult.has_value()) {
|
|
LOG(log_.error()) << "Failed to fetch nodes data";
|
|
isHealthy_ = false;
|
|
return;
|
|
}
|
|
|
|
// Create a new vector here to not have partially parsed data in otherNodesData_
|
|
std::vector<ClioNode> otherNodesData;
|
|
for (auto const& [uuid, nodeDataStr] : expectedResult.value()) {
|
|
if (uuid == *selfData_.uuid) {
|
|
continue;
|
|
}
|
|
|
|
boost::system::error_code errorCode;
|
|
auto const json = boost::json::parse(nodeDataStr, errorCode);
|
|
if (errorCode.failed()) {
|
|
LOG(log_.error()) << "Error parsing json from DB: " << nodeDataStr;
|
|
isHealthy_ = false;
|
|
return;
|
|
}
|
|
|
|
auto expectedNodeData = boost::json::try_value_to<ClioNode>(json);
|
|
if (expectedNodeData.has_error()) {
|
|
LOG(log_.error()) << "Error converting json to ClioNode: " << json;
|
|
isHealthy_ = false;
|
|
return;
|
|
}
|
|
*expectedNodeData->uuid = uuid;
|
|
otherNodesData.push_back(std::move(expectedNodeData).value());
|
|
}
|
|
otherNodesData_ = std::move(otherNodesData);
|
|
nodesInClusterMetric_.set(otherNodesData_.size() + 1);
|
|
isHealthy_ = true;
|
|
}
|
|
|
|
void
|
|
ClusterCommunicationService::doWrite()
|
|
{
|
|
selfData_.updateTime = std::chrono::system_clock::now();
|
|
boost::json::value jsonValue{};
|
|
boost::json::value_from(selfData_, jsonValue);
|
|
backend_->writeNodeMessage(*selfData_.uuid, boost::json::serialize(jsonValue.as_object()));
|
|
}
|
|
|
|
} // namespace cluster
|