2
0
Fork 0
mirror of https://git.asonix.dog/asonix/pict-rs synced 2024-11-09 22:14:59 +00:00

postgres: Rework job & upload notifications (more)

postgres: Add metrics to job push & pop, upload wait
sled: add upload wait metrics
This commit is contained in:
asonix 2023-09-04 16:19:46 -05:00
parent ac9777782a
commit 37448722de
8 changed files with 283 additions and 95 deletions

View file

@ -103,6 +103,7 @@ impl Details {
)) ))
} }
#[tracing::instrument(level = "DEBUG")]
pub(crate) async fn from_store<S: Store>( pub(crate) async fn from_store<S: Store>(
store: &S, store: &S,
identifier: &Arc<str>, identifier: &Arc<str>,

View file

@ -9,6 +9,7 @@ use crate::{
use super::Discovery; use super::Discovery;
#[tracing::instrument(level = "DEBUG", skip_all)]
pub(super) async fn check_reorient( pub(super) async fn check_reorient(
Discovery { Discovery {
input, input,

View file

@ -97,6 +97,7 @@ pub(super) async fn confirm_bytes(
.await .await
} }
#[tracing::instrument(level = "DEBUG", skip(f))]
async fn count_avif_frames<F, Fut>(f: F, timeout: u64) -> Result<u32, MagickError> async fn count_avif_frames<F, Fut>(f: F, timeout: u64) -> Result<u32, MagickError>
where where
F: FnOnce(crate::file::File) -> Fut, F: FnOnce(crate::file::File) -> Fut,
@ -147,6 +148,7 @@ where
Ok(lines) Ok(lines)
} }
#[tracing::instrument(level = "DEBUG", skip(f))]
async fn discover_file<F, Fut>(f: F, timeout: u64) -> Result<Discovery, MagickError> async fn discover_file<F, Fut>(f: F, timeout: u64) -> Result<Discovery, MagickError>
where where
F: FnOnce(crate::file::File) -> Fut, F: FnOnce(crate::file::File) -> Fut,

View file

@ -1,3 +1,9 @@
mod alias;
mod delete_token;
mod hash;
mod metrics;
mod migrate;
use crate::{ use crate::{
config, config,
details::Details, details::Details,
@ -9,10 +15,6 @@ use std::{fmt::Debug, sync::Arc};
use url::Url; use url::Url;
use uuid::Uuid; use uuid::Uuid;
mod alias;
mod delete_token;
mod hash;
mod migrate;
pub(crate) mod postgres; pub(crate) mod postgres;
pub(crate) mod sled; pub(crate) mod sled;

74
src/repo/metrics.rs Normal file
View file

@ -0,0 +1,74 @@
use std::time::Instant;
pub(super) struct PushMetricsGuard {
queue: &'static str,
armed: bool,
}
pub(super) struct PopMetricsGuard {
queue: &'static str,
start: Instant,
armed: bool,
}
pub(super) struct WaitMetricsGuard {
start: Instant,
armed: bool,
}
impl PushMetricsGuard {
pub(super) fn guard(queue: &'static str) -> Self {
Self { queue, armed: true }
}
pub(super) fn disarm(mut self) {
self.armed = false;
}
}
impl PopMetricsGuard {
pub(super) fn guard(queue: &'static str) -> Self {
Self {
queue,
start: Instant::now(),
armed: true,
}
}
pub(super) fn disarm(mut self) {
self.armed = false;
}
}
impl WaitMetricsGuard {
pub(super) fn guard() -> Self {
Self {
start: Instant::now(),
armed: true,
}
}
pub(super) fn disarm(mut self) {
self.armed = false;
}
}
impl Drop for PushMetricsGuard {
fn drop(&mut self) {
metrics::increment_counter!("pict-rs.queue.push", "completed" => (!self.armed).to_string(), "queue" => self.queue);
}
}
impl Drop for PopMetricsGuard {
fn drop(&mut self) {
metrics::histogram!("pict-rs.queue.pop.duration", self.start.elapsed().as_secs_f64(), "completed" => (!self.armed).to_string(), "queue" => self.queue);
metrics::increment_counter!("pict-rs.queue.pop", "completed" => (!self.armed).to_string(), "queue" => self.queue);
}
}
impl Drop for WaitMetricsGuard {
fn drop(&mut self) {
metrics::histogram!("pict-rs.upload.wait.duration", self.start.elapsed().as_secs_f64(), "completed" => (!self.armed).to_string());
metrics::increment_counter!("pict-rs.upload.wait", "completed" => (!self.armed).to_string());
}
}

View file

@ -3,11 +3,12 @@ mod job_status;
mod schema; mod schema;
use std::{ use std::{
collections::{BTreeSet, VecDeque},
sync::{ sync::{
atomic::{AtomicU64, Ordering}, atomic::{AtomicU64, Ordering},
Arc, Arc, Weak,
}, },
time::Duration, time::{Duration, Instant},
}; };
use dashmap::DashMap; use dashmap::DashMap;
@ -35,6 +36,7 @@ use crate::{
use self::job_status::JobStatus; use self::job_status::JobStatus;
use super::{ use super::{
metrics::{PopMetricsGuard, PushMetricsGuard, WaitMetricsGuard},
Alias, AliasAccessRepo, AliasAlreadyExists, AliasRepo, BaseRepo, DeleteToken, DetailsRepo, Alias, AliasAccessRepo, AliasAlreadyExists, AliasRepo, BaseRepo, DeleteToken, DetailsRepo,
FullRepo, Hash, HashAlreadyExists, HashPage, HashRepo, JobId, OrderedHash, ProxyRepo, FullRepo, Hash, HashAlreadyExists, HashPage, HashRepo, JobId, OrderedHash, ProxyRepo,
QueueRepo, RepoError, SettingsRepo, StoreMigrationRepo, UploadId, UploadRepo, UploadResult, QueueRepo, RepoError, SettingsRepo, StoreMigrationRepo, UploadId, UploadRepo, UploadResult,
@ -52,35 +54,24 @@ struct Inner {
health_count: AtomicU64, health_count: AtomicU64,
pool: Pool<AsyncPgConnection>, pool: Pool<AsyncPgConnection>,
queue_notifications: DashMap<String, Arc<Notify>>, queue_notifications: DashMap<String, Arc<Notify>>,
upload_notifier: Notify, upload_notifications: DashMap<UploadId, Weak<Notify>>,
} }
async fn delegate_notifications(receiver: flume::Receiver<Notification>, inner: Arc<Inner>) { struct UploadInterest {
while let Ok(notification) = receiver.recv_async().await { inner: Arc<Inner>,
match notification.channel() { interest: Option<Arc<Notify>>,
"queue_status_channel" => { upload_id: UploadId,
// new job inserted for queue }
let queue_name = notification.payload().to_string();
inner struct JobNotifierState<'a> {
.queue_notifications inner: &'a Inner,
.entry(queue_name) capacity: usize,
.or_insert_with(crate::sync::notify) jobs: BTreeSet<JobId>,
.notify_waiters(); jobs_ordered: VecDeque<JobId>,
} }
"upload_completion_channel" => {
inner.upload_notifier.notify_waiters();
}
channel => {
tracing::info!(
"Unhandled postgres notification: {channel}: {}",
notification.payload()
);
}
}
}
tracing::warn!("Notification delegator shutting down"); struct UploadNotifierState<'a> {
inner: &'a Inner,
} }
#[derive(Debug, thiserror::Error)] #[derive(Debug, thiserror::Error)]
@ -166,7 +157,7 @@ impl PostgresRepo {
health_count: AtomicU64::new(0), health_count: AtomicU64::new(0),
pool, pool,
queue_notifications: DashMap::new(), queue_notifications: DashMap::new(),
upload_notifier: crate::sync::bare_notify(), upload_notifications: DashMap::new(),
}); });
let handle = crate::sync::spawn(delegate_notifications(rx, inner.clone())); let handle = crate::sync::spawn(delegate_notifications(rx, inner.clone()));
@ -184,10 +175,125 @@ impl PostgresRepo {
} }
} }
struct GetConnectionMetricsGuard {
start: Instant,
armed: bool,
}
impl GetConnectionMetricsGuard {
fn guard() -> Self {
GetConnectionMetricsGuard {
start: Instant::now(),
armed: true,
}
}
fn disarm(mut self) {
self.armed = false;
}
}
impl Drop for GetConnectionMetricsGuard {
fn drop(&mut self) {
metrics::increment_counter!("pict-rs.postgres.pool.get.end", "completed" => (!self.armed).to_string());
metrics::histogram!("pict-rs.postgres.pool.get.duration", self.start.elapsed().as_secs_f64(), "completed" => (!self.armed).to_string());
}
}
impl Inner { impl Inner {
#[tracing::instrument(level = "DEBUG", skip(self))] #[tracing::instrument(level = "TRACE", skip(self))]
async fn get_connection(&self) -> Result<Object<AsyncPgConnection>, PostgresError> { async fn get_connection(&self) -> Result<Object<AsyncPgConnection>, PostgresError> {
self.pool.get().await.map_err(PostgresError::Pool) let guard = GetConnectionMetricsGuard::guard();
let res = self.pool.get().await.map_err(PostgresError::Pool);
guard.disarm();
res
}
fn interest(self: &Arc<Self>, upload_id: UploadId) -> UploadInterest {
let notify = crate::sync::notify();
self.upload_notifications
.insert(upload_id, Arc::downgrade(&notify));
UploadInterest {
inner: self.clone(),
interest: Some(notify),
upload_id,
}
}
}
impl UploadInterest {
async fn notified_timeout(&self, timeout: Duration) -> Result<(), tokio::time::error::Elapsed> {
actix_rt::time::timeout(
timeout,
self.interest.as_ref().expect("interest exists").notified(),
)
.await
}
}
impl Drop for UploadInterest {
fn drop(&mut self) {
if let Some(interest) = self.interest.take() {
if Arc::into_inner(interest).is_some() {
self.inner.upload_notifications.remove(&self.upload_id);
}
}
}
}
impl<'a> JobNotifierState<'a> {
fn handle(&mut self, payload: &str) {
let Some((job_id, queue_name)) = payload.split_once(' ') else {
tracing::warn!("Invalid queue payload {payload}");
return;
};
let Ok(job_id) = job_id.parse::<Uuid>().map(JobId) else {
tracing::warn!("Invalid job ID {job_id}");
return;
};
if !self.jobs.insert(job_id) {
// duplicate job
return;
}
self.jobs_ordered.push_back(job_id);
if self.jobs_ordered.len() > self.capacity {
if let Some(job_id) = self.jobs_ordered.pop_front() {
self.jobs.remove(&job_id);
}
}
self.inner
.queue_notifications
.entry(queue_name.to_string())
.or_insert_with(crate::sync::notify)
.notify_one();
metrics::increment_counter!("pict-rs.postgres.job-notifier.notified", "queue" => queue_name.to_string());
}
}
impl<'a> UploadNotifierState<'a> {
fn handle(&self, payload: &str) {
let Ok(upload_id) = payload.parse::<UploadId>() else {
tracing::warn!("Invalid upload id {payload}");
return;
};
if let Some(notifier) = self
.inner
.upload_notifications
.get(&upload_id)
.and_then(|weak| weak.upgrade())
{
notifier.notify_waiters();
metrics::increment_counter!("pict-rs.postgres.upload-notifier.notified");
}
} }
} }
@ -195,6 +301,44 @@ type BoxFuture<'a, T> = std::pin::Pin<Box<dyn std::future::Future<Output = T> +
type ConfigFn = type ConfigFn =
Box<dyn Fn(&str) -> BoxFuture<'_, ConnectionResult<AsyncPgConnection>> + Send + Sync + 'static>; Box<dyn Fn(&str) -> BoxFuture<'_, ConnectionResult<AsyncPgConnection>> + Send + Sync + 'static>;
async fn delegate_notifications(receiver: flume::Receiver<Notification>, inner: Arc<Inner>) {
let parallelism = std::thread::available_parallelism()
.map(|u| u.into())
.unwrap_or(1_usize);
let mut job_notifier_state = JobNotifierState {
inner: &inner,
capacity: parallelism * 8,
jobs: BTreeSet::new(),
jobs_ordered: VecDeque::new(),
};
let upload_notifier_state = UploadNotifierState { inner: &inner };
while let Ok(notification) = receiver.recv_async().await {
metrics::increment_counter!("pict-rs.postgres.notification");
match notification.channel() {
"queue_status_channel" => {
// new job inserted for queue
job_notifier_state.handle(notification.payload());
}
"upload_completion_channel" => {
// new upload finished
upload_notifier_state.handle(notification.payload());
}
channel => {
tracing::info!(
"Unhandled postgres notification: {channel}: {}",
notification.payload()
);
}
}
}
tracing::warn!("Notification delegator shutting down");
}
fn build_handler(sender: flume::Sender<Notification>) -> ConfigFn { fn build_handler(sender: flume::Sender<Notification>) -> ConfigFn {
Box::new( Box::new(
move |config: &str| -> BoxFuture<'_, ConnectionResult<AsyncPgConnection>> { move |config: &str| -> BoxFuture<'_, ConnectionResult<AsyncPgConnection>> {
@ -834,6 +978,8 @@ impl QueueRepo for PostgresRepo {
queue_name: &'static str, queue_name: &'static str,
job_json: serde_json::Value, job_json: serde_json::Value,
) -> Result<JobId, RepoError> { ) -> Result<JobId, RepoError> {
let guard = PushMetricsGuard::guard(queue_name);
use schema::job_queue::dsl::*; use schema::job_queue::dsl::*;
let mut conn = self.get_connection().await?; let mut conn = self.get_connection().await?;
@ -845,6 +991,8 @@ impl QueueRepo for PostgresRepo {
.await .await
.map_err(PostgresError::Diesel)?; .map_err(PostgresError::Diesel)?;
guard.disarm();
Ok(JobId(job_id)) Ok(JobId(job_id))
} }
@ -854,6 +1002,8 @@ impl QueueRepo for PostgresRepo {
queue_name: &'static str, queue_name: &'static str,
worker_id: Uuid, worker_id: Uuid,
) -> Result<(JobId, serde_json::Value), RepoError> { ) -> Result<(JobId, serde_json::Value), RepoError> {
let guard = PopMetricsGuard::guard(queue_name);
use schema::job_queue::dsl::*; use schema::job_queue::dsl::*;
loop { loop {
@ -923,6 +1073,7 @@ impl QueueRepo for PostgresRepo {
}; };
if let Some((job_id, job_json)) = opt { if let Some((job_id, job_json)) = opt {
guard.disarm();
return Ok((JobId(job_id), job_json)); return Ok((JobId(job_id), job_json));
} }
@ -1334,9 +1485,14 @@ impl UploadRepo for PostgresRepo {
#[tracing::instrument(level = "DEBUG", skip(self))] #[tracing::instrument(level = "DEBUG", skip(self))]
async fn wait(&self, upload_id: UploadId) -> Result<UploadResult, RepoError> { async fn wait(&self, upload_id: UploadId) -> Result<UploadResult, RepoError> {
let guard = WaitMetricsGuard::guard();
use schema::uploads::dsl::*; use schema::uploads::dsl::*;
let interest = self.inner.interest(upload_id);
loop { loop {
let interest_future = interest.notified_timeout(Duration::from_secs(5));
let mut conn = self.get_connection().await?; let mut conn = self.get_connection().await?;
diesel::sql_query("LISTEN upload_completion_channel;") diesel::sql_query("LISTEN upload_completion_channel;")
@ -1359,6 +1515,7 @@ impl UploadRepo for PostgresRepo {
serde_json::from_value(upload_result) serde_json::from_value(upload_result)
.map_err(PostgresError::DeserializeUploadResult)?; .map_err(PostgresError::DeserializeUploadResult)?;
guard.disarm();
return Ok(upload_result.into()); return Ok(upload_result.into());
} }
} }
@ -1369,13 +1526,7 @@ impl UploadRepo for PostgresRepo {
drop(conn); drop(conn);
if actix_rt::time::timeout( if interest_future.await.is_ok() {
Duration::from_secs(5),
self.inner.upload_notifier.notified(),
)
.await
.is_ok()
{
tracing::debug!("Notified"); tracing::debug!("Notified");
} else { } else {
tracing::debug!("Timed out"); tracing::debug!("Timed out");

View file

@ -34,7 +34,7 @@ CREATE OR REPLACE FUNCTION queue_status_notify()
RETURNS trigger AS RETURNS trigger AS
$$ $$
BEGIN BEGIN
PERFORM pg_notify('queue_status_channel', NEW.queue::text); PERFORM pg_notify('queue_status_channel', NEW.id::text || ' ' || NEW.queue::text);
RETURN NEW; RETURN NEW;
END; END;
$$ LANGUAGE plpgsql; $$ LANGUAGE plpgsql;

View file

@ -12,17 +12,18 @@ use std::{
atomic::{AtomicU64, Ordering}, atomic::{AtomicU64, Ordering},
Arc, RwLock, Arc, RwLock,
}, },
time::Instant,
}; };
use tokio::sync::Notify; use tokio::sync::Notify;
use url::Url; use url::Url;
use uuid::Uuid; use uuid::Uuid;
use super::{ use super::{
hash::Hash, Alias, AliasAccessRepo, AliasAlreadyExists, AliasRepo, BaseRepo, DeleteToken, hash::Hash,
Details, DetailsRepo, FullRepo, HashAlreadyExists, HashPage, HashRepo, JobId, OrderedHash, metrics::{PopMetricsGuard, PushMetricsGuard, WaitMetricsGuard},
ProxyRepo, QueueRepo, RepoError, SettingsRepo, StoreMigrationRepo, UploadId, UploadRepo, Alias, AliasAccessRepo, AliasAlreadyExists, AliasRepo, BaseRepo, DeleteToken, Details,
UploadResult, VariantAccessRepo, VariantAlreadyExists, DetailsRepo, FullRepo, HashAlreadyExists, HashPage, HashRepo, JobId, OrderedHash, ProxyRepo,
QueueRepo, RepoError, SettingsRepo, StoreMigrationRepo, UploadId, UploadRepo, UploadResult,
VariantAccessRepo, VariantAlreadyExists,
}; };
macro_rules! b { macro_rules! b {
@ -490,54 +491,6 @@ impl From<InnerUploadResult> for UploadResult {
} }
} }
struct PushMetricsGuard {
queue: &'static str,
armed: bool,
}
struct PopMetricsGuard {
queue: &'static str,
start: Instant,
armed: bool,
}
impl PushMetricsGuard {
fn guard(queue: &'static str) -> Self {
Self { queue, armed: true }
}
fn disarm(mut self) {
self.armed = false;
}
}
impl PopMetricsGuard {
fn guard(queue: &'static str) -> Self {
Self {
queue,
start: Instant::now(),
armed: true,
}
}
fn disarm(mut self) {
self.armed = false;
}
}
impl Drop for PushMetricsGuard {
fn drop(&mut self) {
metrics::increment_counter!("pict-rs.queue.push", "completed" => (!self.armed).to_string(), "queue" => self.queue);
}
}
impl Drop for PopMetricsGuard {
fn drop(&mut self) {
metrics::histogram!("pict-rs.queue.pop.duration", self.start.elapsed().as_secs_f64(), "completed" => (!self.armed).to_string(), "queue" => self.queue);
metrics::increment_counter!("pict-rs.queue.pop", "completed" => (!self.armed).to_string(), "queue" => self.queue);
}
}
#[async_trait::async_trait(?Send)] #[async_trait::async_trait(?Send)]
impl UploadRepo for SledRepo { impl UploadRepo for SledRepo {
#[tracing::instrument(level = "trace", skip(self))] #[tracing::instrument(level = "trace", skip(self))]
@ -551,6 +504,7 @@ impl UploadRepo for SledRepo {
#[tracing::instrument(skip(self))] #[tracing::instrument(skip(self))]
async fn wait(&self, upload_id: UploadId) -> Result<UploadResult, RepoError> { async fn wait(&self, upload_id: UploadId) -> Result<UploadResult, RepoError> {
let guard = WaitMetricsGuard::guard();
let mut subscriber = self.uploads.watch_prefix(upload_id.as_bytes()); let mut subscriber = self.uploads.watch_prefix(upload_id.as_bytes());
let bytes = upload_id.as_bytes().to_vec(); let bytes = upload_id.as_bytes().to_vec();
@ -560,6 +514,7 @@ impl UploadRepo for SledRepo {
if bytes != b"1" { if bytes != b"1" {
let result: InnerUploadResult = let result: InnerUploadResult =
serde_json::from_slice(&bytes).map_err(SledError::UploadResult)?; serde_json::from_slice(&bytes).map_err(SledError::UploadResult)?;
guard.disarm();
return Ok(result.into()); return Ok(result.into());
} }
} else { } else {
@ -575,6 +530,8 @@ impl UploadRepo for SledRepo {
if value != b"1" { if value != b"1" {
let result: InnerUploadResult = let result: InnerUploadResult =
serde_json::from_slice(&value).map_err(SledError::UploadResult)?; serde_json::from_slice(&value).map_err(SledError::UploadResult)?;
guard.disarm();
return Ok(result.into()); return Ok(result.into());
} }
} }