mirror of
https://github.com/LemmyNet/lemmy.git
synced 2024-10-01 01:36:12 -04:00
Send activities in parallel (fixes #4529)
This commit is contained in:
parent
44e9622b71
commit
748b96e513
@ -16,7 +16,7 @@ use once_cell::sync::Lazy;
|
|||||||
use std::{
|
use std::{
|
||||||
cmp::Ordering,
|
cmp::Ordering,
|
||||||
collections::BinaryHeap,
|
collections::BinaryHeap,
|
||||||
sync::{Arc, Mutex},
|
sync::{Arc, RwLock},
|
||||||
thread::available_parallelism,
|
thread::available_parallelism,
|
||||||
time::Duration,
|
time::Duration,
|
||||||
};
|
};
|
||||||
@ -38,26 +38,26 @@ pub async fn shared_inbox(
|
|||||||
request.method().clone(),
|
request.method().clone(),
|
||||||
request.uri().clone(),
|
request.uri().clone(),
|
||||||
);
|
);
|
||||||
ACTIVITY_QUEUE.lock().unwrap().push(InboxActivity {
|
ACTIVITY_QUEUE.write().unwrap().push(InboxActivity {
|
||||||
request_parts,
|
request_parts,
|
||||||
bytes,
|
bytes,
|
||||||
published,
|
published,
|
||||||
});
|
});
|
||||||
|
Ok(HttpResponse::Ok().finish())
|
||||||
}
|
}
|
||||||
None => {
|
None => {
|
||||||
// no timestamp included, process immediately
|
// no timestamp included, process immediately
|
||||||
receive_activity::<SharedInboxActivities, UserOrCommunity, LemmyContext>(
|
receive_activity::<SharedInboxActivities, UserOrCommunity, LemmyContext>(
|
||||||
request, bytes, &data,
|
request, bytes, &data,
|
||||||
)
|
)
|
||||||
.await?;
|
.await
|
||||||
}
|
}
|
||||||
};
|
}
|
||||||
Ok(HttpResponse::Ok().finish())
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Queue of incoming activities, ordered by oldest published first
|
/// Queue of incoming activities, ordered by oldest published first
|
||||||
static ACTIVITY_QUEUE: Lazy<Arc<Mutex<BinaryHeap<InboxActivity>>>> =
|
static ACTIVITY_QUEUE: Lazy<Arc<RwLock<BinaryHeap<InboxActivity>>>> =
|
||||||
Lazy::new(|| Arc::new(Mutex::new(BinaryHeap::new())));
|
Lazy::new(|| Arc::new(RwLock::new(BinaryHeap::new())));
|
||||||
|
|
||||||
/// Minimum age of an activity before it gets processed. This ensures that an activity which was
|
/// Minimum age of an activity before it gets processed. This ensures that an activity which was
|
||||||
/// delayed still gets processed in correct order.
|
/// delayed still gets processed in correct order.
|
||||||
@ -100,11 +100,11 @@ pub fn handle_received_activities(
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn peek_queue_timestamp() -> Option<DateTime<Utc>> {
|
fn peek_queue_timestamp() -> Option<DateTime<Utc>> {
|
||||||
ACTIVITY_QUEUE.lock().unwrap().peek().map(|i| i.published)
|
ACTIVITY_QUEUE.read().unwrap().peek().map(|i| i.published)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn pop_queue<'a>() -> Option<InboxActivity> {
|
fn pop_queue<'a>() -> Option<InboxActivity> {
|
||||||
ACTIVITY_QUEUE.lock().unwrap().pop()
|
ACTIVITY_QUEUE.write().unwrap().pop()
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, Debug)]
|
#[derive(Clone, Debug)]
|
||||||
@ -156,7 +156,7 @@ mod tests {
|
|||||||
bytes: Default::default(),
|
bytes: Default::default(),
|
||||||
published: Local::now().into(),
|
published: Local::now().into(),
|
||||||
};
|
};
|
||||||
let mut lock = ACTIVITY_QUEUE.lock().unwrap();
|
let mut lock = ACTIVITY_QUEUE.write().unwrap();
|
||||||
|
|
||||||
// insert in wrong order
|
// insert in wrong order
|
||||||
lock.push(activity3.clone());
|
lock.push(activity3.clone());
|
||||||
|
@ -51,7 +51,7 @@ impl ActivitySendTargets {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(PartialEq, Eq, Debug)]
|
#[derive(PartialEq, Eq, Debug, Clone)]
|
||||||
#[cfg_attr(feature = "full", derive(Queryable, Selectable, Identifiable))]
|
#[cfg_attr(feature = "full", derive(Queryable, Selectable, Identifiable))]
|
||||||
#[cfg_attr(feature = "full", diesel(check_for_backend(diesel::pg::Pg)))]
|
#[cfg_attr(feature = "full", diesel(check_for_backend(diesel::pg::Pg)))]
|
||||||
#[cfg_attr(feature = "full", diesel(table_name = sent_activity))]
|
#[cfg_attr(feature = "full", diesel(table_name = sent_activity))]
|
||||||
|
@ -30,9 +30,10 @@ use reqwest::Url;
|
|||||||
use std::{
|
use std::{
|
||||||
collections::{HashMap, HashSet},
|
collections::{HashMap, HashSet},
|
||||||
ops::{Add, Deref},
|
ops::{Add, Deref},
|
||||||
|
sync::{Arc, RwLock},
|
||||||
time::Duration,
|
time::Duration,
|
||||||
};
|
};
|
||||||
use tokio::{sync::mpsc::UnboundedSender, time::sleep};
|
use tokio::{spawn, sync::mpsc::UnboundedSender, time::sleep};
|
||||||
use tokio_util::sync::CancellationToken;
|
use tokio_util::sync::CancellationToken;
|
||||||
|
|
||||||
/// Check whether to save state to db every n sends if there's no failures (during failures state is saved after every attempt)
|
/// Check whether to save state to db every n sends if there's no failures (during failures state is saved after every attempt)
|
||||||
@ -57,8 +58,11 @@ static FOLLOW_ADDITIONS_RECHECK_DELAY: Lazy<chrono::TimeDelta> = Lazy::new(|| {
|
|||||||
/// This is expected to happen pretty rarely and updating it in a timely manner is not too important.
|
/// This is expected to happen pretty rarely and updating it in a timely manner is not too important.
|
||||||
static FOLLOW_REMOVALS_RECHECK_DELAY: Lazy<chrono::TimeDelta> =
|
static FOLLOW_REMOVALS_RECHECK_DELAY: Lazy<chrono::TimeDelta> =
|
||||||
Lazy::new(|| chrono::TimeDelta::try_hours(1).expect("TimeDelta out of bounds"));
|
Lazy::new(|| chrono::TimeDelta::try_hours(1).expect("TimeDelta out of bounds"));
|
||||||
|
|
||||||
|
const MAX_INFLIGHT_REQUESTS: u8 = 5;
|
||||||
|
|
||||||
pub(crate) struct InstanceWorker {
|
pub(crate) struct InstanceWorker {
|
||||||
instance: Instance,
|
instance: Arc<RwLock<Instance>>,
|
||||||
// load site lazily because if an instance is first seen due to being on allowlist,
|
// load site lazily because if an instance is first seen due to being on allowlist,
|
||||||
// the corresponding row in `site` may not exist yet since that is only added once
|
// the corresponding row in `site` may not exist yet since that is only added once
|
||||||
// `fetch_instance_actor_for_object` is called.
|
// `fetch_instance_actor_for_object` is called.
|
||||||
@ -68,14 +72,26 @@ pub(crate) struct InstanceWorker {
|
|||||||
followed_communities: HashMap<CommunityId, HashSet<Url>>,
|
followed_communities: HashMap<CommunityId, HashSet<Url>>,
|
||||||
stop: CancellationToken,
|
stop: CancellationToken,
|
||||||
context: Data<LemmyContext>,
|
context: Data<LemmyContext>,
|
||||||
stats_sender: UnboundedSender<(String, FederationQueueState)>,
|
|
||||||
last_full_communities_fetch: DateTime<Utc>,
|
last_full_communities_fetch: DateTime<Utc>,
|
||||||
last_incremental_communities_fetch: DateTime<Utc>,
|
last_incremental_communities_fetch: DateTime<Utc>,
|
||||||
|
stats: Arc<RwLock<InstanceStats>>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Clone)]
|
||||||
|
struct InstanceStats {
|
||||||
|
stats_sender: UnboundedSender<(String, FederationQueueState)>,
|
||||||
state: FederationQueueState,
|
state: FederationQueueState,
|
||||||
last_state_insert: DateTime<Utc>,
|
last_state_insert: DateTime<Utc>,
|
||||||
|
inflight_requests: u8,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl InstanceWorker {
|
impl InstanceWorker {
|
||||||
|
fn stats(&self) -> InstanceStats {
|
||||||
|
self.stats.read().unwrap().clone()
|
||||||
|
}
|
||||||
|
fn instance(&self) -> Instance {
|
||||||
|
self.instance.read().unwrap().clone()
|
||||||
|
}
|
||||||
pub(crate) async fn init_and_loop(
|
pub(crate) async fn init_and_loop(
|
||||||
instance: Instance,
|
instance: Instance,
|
||||||
context: Data<LemmyContext>,
|
context: Data<LemmyContext>,
|
||||||
@ -85,17 +101,20 @@ impl InstanceWorker {
|
|||||||
) -> Result<(), anyhow::Error> {
|
) -> Result<(), anyhow::Error> {
|
||||||
let state = FederationQueueState::load(pool, instance.id).await?;
|
let state = FederationQueueState::load(pool, instance.id).await?;
|
||||||
let mut worker = InstanceWorker {
|
let mut worker = InstanceWorker {
|
||||||
instance,
|
instance: Arc::new(RwLock::new(instance)),
|
||||||
site_loaded: false,
|
site_loaded: false,
|
||||||
site: None,
|
site: None,
|
||||||
followed_communities: HashMap::new(),
|
followed_communities: HashMap::new(),
|
||||||
stop,
|
stop,
|
||||||
context,
|
context,
|
||||||
stats_sender,
|
|
||||||
last_full_communities_fetch: Utc.timestamp_nanos(0),
|
last_full_communities_fetch: Utc.timestamp_nanos(0),
|
||||||
last_incremental_communities_fetch: Utc.timestamp_nanos(0),
|
last_incremental_communities_fetch: Utc.timestamp_nanos(0),
|
||||||
state,
|
stats: Arc::new(RwLock::new(InstanceStats {
|
||||||
last_state_insert: Utc.timestamp_nanos(0),
|
stats_sender,
|
||||||
|
state,
|
||||||
|
last_state_insert: Utc.timestamp_nanos(0),
|
||||||
|
inflight_requests: 0,
|
||||||
|
})),
|
||||||
};
|
};
|
||||||
worker.loop_until_stopped(pool).await
|
worker.loop_until_stopped(pool).await
|
||||||
}
|
}
|
||||||
@ -114,25 +133,26 @@ impl InstanceWorker {
|
|||||||
if self.stop.is_cancelled() {
|
if self.stop.is_cancelled() {
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
if (Utc::now() - self.last_state_insert) > save_state_every {
|
if (Utc::now() - self.stats().last_state_insert) > save_state_every {
|
||||||
self.save_and_send_state(pool).await?;
|
save_and_send_state(self.stats.clone(), &self.instance(), pool).await?;
|
||||||
}
|
}
|
||||||
self.update_communities(pool).await?;
|
self.update_communities(pool).await?;
|
||||||
}
|
}
|
||||||
// final update of state in db
|
// final update of state in db
|
||||||
self.save_and_send_state(pool).await?;
|
save_and_send_state(self.stats.clone(), &self.instance(), pool).await?;
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn initial_fail_sleep(&mut self) -> Result<()> {
|
async fn initial_fail_sleep(&mut self) -> Result<()> {
|
||||||
// before starting queue, sleep remaining duration if last request failed
|
// before starting queue, sleep remaining duration if last request failed
|
||||||
if self.state.fail_count > 0 {
|
let stats = self.stats();
|
||||||
let last_retry = self
|
if stats.state.fail_count > 0 {
|
||||||
|
let last_retry = stats
|
||||||
.state
|
.state
|
||||||
.last_retry
|
.last_retry
|
||||||
.context("impossible: if fail count set last retry also set")?;
|
.context("impossible: if fail count set last retry also set")?;
|
||||||
let elapsed = (Utc::now() - last_retry).to_std()?;
|
let elapsed = (Utc::now() - last_retry).to_std()?;
|
||||||
let required = federate_retry_sleep_duration(self.state.fail_count);
|
let required = federate_retry_sleep_duration(stats.state.fail_count);
|
||||||
if elapsed >= required {
|
if elapsed >= required {
|
||||||
return Ok(());
|
return Ok(());
|
||||||
}
|
}
|
||||||
@ -147,14 +167,16 @@ impl InstanceWorker {
|
|||||||
/// send out a batch of CHECK_SAVE_STATE_EVERY_IT activities
|
/// send out a batch of CHECK_SAVE_STATE_EVERY_IT activities
|
||||||
async fn loop_batch(&mut self, pool: &mut DbPool<'_>) -> Result<()> {
|
async fn loop_batch(&mut self, pool: &mut DbPool<'_>) -> Result<()> {
|
||||||
let latest_id = get_latest_activity_id(pool).await?;
|
let latest_id = get_latest_activity_id(pool).await?;
|
||||||
let mut id = if let Some(id) = self.state.last_successful_id {
|
let mut id = if let Some(id) = self.stats().state.last_successful_id {
|
||||||
id
|
id
|
||||||
} else {
|
} else {
|
||||||
// this is the initial creation (instance first seen) of the federation queue for this instance
|
// this is the initial creation (instance first seen) of the federation queue for this instance
|
||||||
// skip all past activities:
|
// skip all past activities:
|
||||||
self.state.last_successful_id = Some(latest_id);
|
{
|
||||||
|
self.stats.write().unwrap().state.last_successful_id = Some(latest_id);
|
||||||
|
}
|
||||||
// save here to ensure it's not read as 0 again later if no activities have happened
|
// save here to ensure it's not read as 0 again later if no activities have happened
|
||||||
self.save_and_send_state(pool).await?;
|
save_and_send_state(self.stats.clone(), &self.instance(), pool).await?;
|
||||||
latest_id
|
latest_id
|
||||||
};
|
};
|
||||||
if id >= latest_id {
|
if id >= latest_id {
|
||||||
@ -170,30 +192,20 @@ impl InstanceWorker {
|
|||||||
&& processed_activities < CHECK_SAVE_STATE_EVERY_IT
|
&& processed_activities < CHECK_SAVE_STATE_EVERY_IT
|
||||||
&& !self.stop.is_cancelled()
|
&& !self.stop.is_cancelled()
|
||||||
{
|
{
|
||||||
|
while self.stats().inflight_requests >= MAX_INFLIGHT_REQUESTS {
|
||||||
|
sleep(Duration::from_millis(100)).await;
|
||||||
|
}
|
||||||
id = ActivityId(id.0 + 1);
|
id = ActivityId(id.0 + 1);
|
||||||
processed_activities += 1;
|
processed_activities += 1;
|
||||||
let Some(ele) = get_activity_cached(pool, id)
|
let Some(ele) = get_activity_cached(pool, id)
|
||||||
.await
|
.await
|
||||||
.context("failed reading activity from db")?
|
.context("failed reading activity from db")?
|
||||||
else {
|
else {
|
||||||
tracing::debug!("{}: {:?} does not exist", self.instance.domain, id);
|
tracing::debug!("{}: {:?} does not exist", self.instance().domain, id);
|
||||||
self.state.last_successful_id = Some(id);
|
self.stats.write().unwrap().state.last_successful_id = Some(id);
|
||||||
continue;
|
continue;
|
||||||
};
|
};
|
||||||
if let Err(e) = self.send_retry_loop(pool, &ele.0, &ele.1).await {
|
self.send_retry_loop(pool, ele.0.clone(), &ele.1).await?
|
||||||
tracing::warn!(
|
|
||||||
"sending {} errored internally, skipping activity: {:?}",
|
|
||||||
ele.0.ap_id,
|
|
||||||
e
|
|
||||||
);
|
|
||||||
}
|
|
||||||
if self.stop.is_cancelled() {
|
|
||||||
return Ok(());
|
|
||||||
}
|
|
||||||
// send success!
|
|
||||||
self.state.last_successful_id = Some(id);
|
|
||||||
self.state.last_successful_published_time = Some(ele.0.published);
|
|
||||||
self.state.fail_count = 0;
|
|
||||||
}
|
}
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
@ -203,17 +215,34 @@ impl InstanceWorker {
|
|||||||
async fn send_retry_loop(
|
async fn send_retry_loop(
|
||||||
&mut self,
|
&mut self,
|
||||||
pool: &mut DbPool<'_>,
|
pool: &mut DbPool<'_>,
|
||||||
activity: &SentActivity,
|
activity: SentActivity,
|
||||||
object: &SharedInboxActivities,
|
object: &SharedInboxActivities,
|
||||||
) -> Result<()> {
|
) -> Result<()> {
|
||||||
|
let stats = self.stats();
|
||||||
|
let retry_delay: Duration = federate_retry_sleep_duration(stats.state.fail_count);
|
||||||
|
tracing::info!(
|
||||||
|
"{}: retrying {:?} attempt {} with delay {retry_delay:.2?}",
|
||||||
|
self.instance().domain,
|
||||||
|
activity.id,
|
||||||
|
stats.state.fail_count
|
||||||
|
);
|
||||||
|
tokio::select! {
|
||||||
|
() = sleep(retry_delay) => {},
|
||||||
|
() = self.stop.cancelled() => {
|
||||||
|
// save state to db and exit
|
||||||
|
return Ok(());
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
let inbox_urls = self
|
let inbox_urls = self
|
||||||
.get_inbox_urls(pool, activity)
|
.get_inbox_urls(pool, &activity)
|
||||||
.await
|
.await
|
||||||
.context("failed figuring out inbox urls")?;
|
.context("failed figuring out inbox urls")?;
|
||||||
if inbox_urls.is_empty() {
|
if inbox_urls.is_empty() {
|
||||||
tracing::debug!("{}: {:?} no inboxes", self.instance.domain, activity.id);
|
tracing::debug!("{}: {:?} no inboxes", self.instance().domain, activity.id);
|
||||||
self.state.last_successful_id = Some(activity.id);
|
let mut stats = self.stats.write().unwrap();
|
||||||
self.state.last_successful_published_time = Some(activity.published);
|
stats.state.last_successful_id = Some(activity.id);
|
||||||
|
stats.state.last_successful_published_time = Some(activity.published);
|
||||||
return Ok(());
|
return Ok(());
|
||||||
}
|
}
|
||||||
let Some(actor_apub_id) = &activity.actor_apub_id else {
|
let Some(actor_apub_id) = &activity.actor_apub_id else {
|
||||||
@ -225,43 +254,63 @@ impl InstanceWorker {
|
|||||||
|
|
||||||
let object = WithContext::new(object.clone(), FEDERATION_CONTEXT.deref().clone());
|
let object = WithContext::new(object.clone(), FEDERATION_CONTEXT.deref().clone());
|
||||||
let inbox_urls = inbox_urls.into_iter().collect();
|
let inbox_urls = inbox_urls.into_iter().collect();
|
||||||
let requests =
|
let context = self.context.reset_request_count();
|
||||||
SendActivityTask::prepare(&object, actor.as_ref(), inbox_urls, &self.context).await?;
|
let stats = self.stats.clone();
|
||||||
for task in requests {
|
let instance = self.instance();
|
||||||
|
let write_instance = self.instance.clone();
|
||||||
|
spawn(async move {
|
||||||
|
{
|
||||||
|
stats.write().unwrap().inflight_requests += 1;
|
||||||
|
}
|
||||||
|
let requests = SendActivityTask::prepare(&object, actor.as_ref(), inbox_urls, &context)
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
// usually only one due to shared inbox
|
// usually only one due to shared inbox
|
||||||
tracing::debug!("sending out {}", task);
|
for task in requests {
|
||||||
while let Err(e) = task.sign_and_send(&self.context).await {
|
tracing::debug!("sending out {}", task);
|
||||||
self.state.fail_count += 1;
|
let res = task.sign_and_send(&context).await;
|
||||||
self.state.last_retry = Some(Utc::now());
|
match res {
|
||||||
let retry_delay: Duration = federate_retry_sleep_duration(self.state.fail_count);
|
Ok(_) => {
|
||||||
tracing::info!(
|
// send success!
|
||||||
"{}: retrying {:?} attempt {} with delay {retry_delay:.2?}. ({e})",
|
{
|
||||||
self.instance.domain,
|
let mut stats_ = stats.write().unwrap();
|
||||||
activity.id,
|
stats_.inflight_requests -= 1;
|
||||||
self.state.fail_count
|
stats_.state.last_successful_id = Some(activity.id);
|
||||||
);
|
stats_.state.last_successful_published_time = Some(activity.published);
|
||||||
self.save_and_send_state(pool).await?;
|
stats_.state.fail_count = 0;
|
||||||
tokio::select! {
|
}
|
||||||
() = sleep(retry_delay) => {},
|
|
||||||
() = self.stop.cancelled() => {
|
// mark instance as alive if it hasn't been updated in a while.
|
||||||
// save state to db and exit
|
let updated = instance.updated.unwrap_or(instance.published);
|
||||||
return Ok(());
|
if updated.add(Days::new(1)) < Utc::now() {
|
||||||
|
{
|
||||||
|
write_instance.write().unwrap().updated = Some(Utc::now());
|
||||||
|
}
|
||||||
|
|
||||||
|
let form = InstanceForm::builder()
|
||||||
|
.domain(instance.domain.clone())
|
||||||
|
.updated(Some(naive_now()))
|
||||||
|
.build();
|
||||||
|
Instance::update(&mut context.pool(), instance.id, form)
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Err(e) => {
|
||||||
|
tracing::info!("{} send failed: {e}", instance.domain);
|
||||||
|
{
|
||||||
|
let mut stats_ = stats.write().unwrap();
|
||||||
|
stats_.inflight_requests -= 1;
|
||||||
|
stats_.state.fail_count += 1;
|
||||||
|
stats_.state.last_retry = Some(Utc::now());
|
||||||
|
}
|
||||||
|
save_and_send_state(stats.clone(), &instance, &mut context.pool())
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
});
|
||||||
// Activity send successful, mark instance as alive if it hasn't been updated in a while.
|
|
||||||
let updated = self.instance.updated.unwrap_or(self.instance.published);
|
|
||||||
if updated.add(Days::new(1)) < Utc::now() {
|
|
||||||
self.instance.updated = Some(Utc::now());
|
|
||||||
|
|
||||||
let form = InstanceForm::builder()
|
|
||||||
.domain(self.instance.domain.clone())
|
|
||||||
.updated(Some(naive_now()))
|
|
||||||
.build();
|
|
||||||
Instance::update(pool, self.instance.id, form).await?;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -278,7 +327,7 @@ impl InstanceWorker {
|
|||||||
|
|
||||||
if activity.send_all_instances {
|
if activity.send_all_instances {
|
||||||
if !self.site_loaded {
|
if !self.site_loaded {
|
||||||
self.site = Site::read_from_instance_id(pool, self.instance.id).await?;
|
self.site = Site::read_from_instance_id(pool, self.instance().id).await?;
|
||||||
self.site_loaded = true;
|
self.site_loaded = true;
|
||||||
}
|
}
|
||||||
if let Some(site) = &self.site {
|
if let Some(site) = &self.site {
|
||||||
@ -296,7 +345,7 @@ impl InstanceWorker {
|
|||||||
.send_inboxes
|
.send_inboxes
|
||||||
.iter()
|
.iter()
|
||||||
.filter_map(std::option::Option::as_ref)
|
.filter_map(std::option::Option::as_ref)
|
||||||
.filter(|&u| (u.domain() == Some(&self.instance.domain)))
|
.filter(|&u| (u.domain() == Some(&self.instance().domain)))
|
||||||
.map(|u| u.inner().clone()),
|
.map(|u| u.inner().clone()),
|
||||||
);
|
);
|
||||||
Ok(inbox_urls)
|
Ok(inbox_urls)
|
||||||
@ -306,7 +355,7 @@ impl InstanceWorker {
|
|||||||
if (Utc::now() - self.last_full_communities_fetch) > *FOLLOW_REMOVALS_RECHECK_DELAY {
|
if (Utc::now() - self.last_full_communities_fetch) > *FOLLOW_REMOVALS_RECHECK_DELAY {
|
||||||
// process removals every hour
|
// process removals every hour
|
||||||
(self.followed_communities, self.last_full_communities_fetch) = self
|
(self.followed_communities, self.last_full_communities_fetch) = self
|
||||||
.get_communities(pool, self.instance.id, Utc.timestamp_nanos(0))
|
.get_communities(pool, self.instance().id, Utc.timestamp_nanos(0))
|
||||||
.await?;
|
.await?;
|
||||||
self.last_incremental_communities_fetch = self.last_full_communities_fetch;
|
self.last_incremental_communities_fetch = self.last_full_communities_fetch;
|
||||||
}
|
}
|
||||||
@ -315,7 +364,7 @@ impl InstanceWorker {
|
|||||||
let (news, time) = self
|
let (news, time) = self
|
||||||
.get_communities(
|
.get_communities(
|
||||||
pool,
|
pool,
|
||||||
self.instance.id,
|
self.instance().id,
|
||||||
self.last_incremental_communities_fetch,
|
self.last_incremental_communities_fetch,
|
||||||
)
|
)
|
||||||
.await?;
|
.await?;
|
||||||
@ -345,12 +394,21 @@ impl InstanceWorker {
|
|||||||
new_last_fetch,
|
new_last_fetch,
|
||||||
))
|
))
|
||||||
}
|
}
|
||||||
async fn save_and_send_state(&mut self, pool: &mut DbPool<'_>) -> Result<()> {
|
}
|
||||||
self.last_state_insert = Utc::now();
|
|
||||||
FederationQueueState::upsert(pool, &self.state).await?;
|
async fn save_and_send_state(
|
||||||
self
|
stats: Arc<RwLock<InstanceStats>>,
|
||||||
.stats_sender
|
instance: &Instance,
|
||||||
.send((self.instance.domain.clone(), self.state.clone()))?;
|
pool: &mut DbPool<'_>,
|
||||||
Ok(())
|
) -> Result<()> {
|
||||||
}
|
let stats = {
|
||||||
|
let mut lock = stats.write().unwrap();
|
||||||
|
lock.last_state_insert = Utc::now();
|
||||||
|
lock.clone()
|
||||||
|
};
|
||||||
|
FederationQueueState::upsert(pool, &stats.state).await?;
|
||||||
|
stats
|
||||||
|
.stats_sender
|
||||||
|
.send((instance.domain.clone(), stats.state))?;
|
||||||
|
Ok(())
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user