2023-09-02 18:52:15 +00:00
|
|
|
mod embedded;
|
2023-09-03 17:47:06 +00:00
|
|
|
mod job_status;
|
2023-09-02 18:35:30 +00:00
|
|
|
mod schema;
|
2023-09-02 16:52:55 +00:00
|
|
|
|
2023-09-03 21:59:41 +00:00
|
|
|
use std::{
|
2023-09-04 21:19:46 +00:00
|
|
|
collections::{BTreeSet, VecDeque},
|
2023-09-03 21:59:41 +00:00
|
|
|
sync::{
|
|
|
|
atomic::{AtomicU64, Ordering},
|
2023-09-04 21:19:46 +00:00
|
|
|
Arc, Weak,
|
2023-09-03 21:59:41 +00:00
|
|
|
},
|
2023-09-04 21:19:46 +00:00
|
|
|
time::{Duration, Instant},
|
2023-09-03 21:59:41 +00:00
|
|
|
};
|
2023-09-02 23:30:45 +00:00
|
|
|
|
2023-09-03 21:59:41 +00:00
|
|
|
use dashmap::DashMap;
|
2023-09-05 02:51:27 +00:00
|
|
|
use deadpool::managed::Hook;
|
2023-09-02 18:52:15 +00:00
|
|
|
use diesel::prelude::*;
|
2023-09-02 16:52:55 +00:00
|
|
|
use diesel_async::{
|
|
|
|
pooled_connection::{
|
2023-09-04 01:05:29 +00:00
|
|
|
deadpool::{BuildError, Object, Pool, PoolError},
|
2023-09-03 01:13:32 +00:00
|
|
|
AsyncDieselConnectionManager, ManagerConfig,
|
2023-09-02 16:52:55 +00:00
|
|
|
},
|
2023-09-03 00:13:00 +00:00
|
|
|
AsyncConnection, AsyncPgConnection, RunQueryDsl,
|
2023-09-02 16:52:55 +00:00
|
|
|
};
|
2023-09-11 00:08:01 +00:00
|
|
|
use futures_core::Stream;
|
2023-09-03 17:47:06 +00:00
|
|
|
use tokio::sync::Notify;
|
2023-09-04 02:30:47 +00:00
|
|
|
use tokio_postgres::{tls::NoTlsStream, AsyncMessage, Connection, NoTls, Notification, Socket};
|
|
|
|
use tracing::Instrument;
|
2023-09-02 16:52:55 +00:00
|
|
|
use url::Url;
|
2023-09-03 17:47:06 +00:00
|
|
|
use uuid::Uuid;
|
2023-09-02 16:52:55 +00:00
|
|
|
|
2023-09-03 21:59:41 +00:00
|
|
|
use crate::{
|
|
|
|
details::Details,
|
|
|
|
error_code::{ErrorCode, OwnedErrorCode},
|
2023-09-11 00:08:01 +00:00
|
|
|
future::{WithMetrics, WithTimeout},
|
2023-09-03 21:59:41 +00:00
|
|
|
serde_str::Serde,
|
|
|
|
stream::LocalBoxStream,
|
|
|
|
};
|
2023-09-03 17:47:06 +00:00
|
|
|
|
|
|
|
use self::job_status::JobStatus;
|
2023-09-02 23:30:45 +00:00
|
|
|
|
|
|
|
use super::{
|
2023-09-04 21:19:46 +00:00
|
|
|
metrics::{PopMetricsGuard, PushMetricsGuard, WaitMetricsGuard},
|
2023-09-03 21:59:41 +00:00
|
|
|
Alias, AliasAccessRepo, AliasAlreadyExists, AliasRepo, BaseRepo, DeleteToken, DetailsRepo,
|
|
|
|
FullRepo, Hash, HashAlreadyExists, HashPage, HashRepo, JobId, OrderedHash, ProxyRepo,
|
|
|
|
QueueRepo, RepoError, SettingsRepo, StoreMigrationRepo, UploadId, UploadRepo, UploadResult,
|
|
|
|
VariantAccessRepo, VariantAlreadyExists,
|
2023-09-02 23:30:45 +00:00
|
|
|
};
|
2023-09-02 16:52:55 +00:00
|
|
|
|
|
|
|
#[derive(Clone)]
|
|
|
|
pub(crate) struct PostgresRepo {
|
2023-09-03 17:47:06 +00:00
|
|
|
inner: Arc<Inner>,
|
2023-09-03 21:59:41 +00:00
|
|
|
#[allow(dead_code)]
|
2023-09-24 20:45:10 +00:00
|
|
|
notifications: Arc<actix_web::rt::task::JoinHandle<()>>,
|
2023-09-03 17:47:06 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
struct Inner {
|
2023-09-03 21:59:41 +00:00
|
|
|
health_count: AtomicU64,
|
2023-09-02 16:52:55 +00:00
|
|
|
pool: Pool<AsyncPgConnection>,
|
2023-09-03 17:47:06 +00:00
|
|
|
queue_notifications: DashMap<String, Arc<Notify>>,
|
2023-09-04 21:19:46 +00:00
|
|
|
upload_notifications: DashMap<UploadId, Weak<Notify>>,
|
2023-09-03 17:47:06 +00:00
|
|
|
}
|
|
|
|
|
2023-09-04 21:19:46 +00:00
|
|
|
struct UploadInterest {
|
|
|
|
inner: Arc<Inner>,
|
|
|
|
interest: Option<Arc<Notify>>,
|
|
|
|
upload_id: UploadId,
|
|
|
|
}
|
2023-09-03 17:47:06 +00:00
|
|
|
|
2023-09-04 21:19:46 +00:00
|
|
|
struct JobNotifierState<'a> {
|
|
|
|
inner: &'a Inner,
|
|
|
|
capacity: usize,
|
|
|
|
jobs: BTreeSet<JobId>,
|
|
|
|
jobs_ordered: VecDeque<JobId>,
|
|
|
|
}
|
2023-09-03 17:47:06 +00:00
|
|
|
|
2023-09-04 21:19:46 +00:00
|
|
|
struct UploadNotifierState<'a> {
|
|
|
|
inner: &'a Inner,
|
2023-09-02 16:52:55 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#[derive(Debug, thiserror::Error)]
|
|
|
|
pub(crate) enum ConnectPostgresError {
|
|
|
|
#[error("Failed to connect to postgres for migrations")]
|
|
|
|
ConnectForMigration(#[source] tokio_postgres::Error),
|
|
|
|
|
|
|
|
#[error("Failed to run migrations")]
|
|
|
|
Migration(#[source] refinery::Error),
|
|
|
|
|
|
|
|
#[error("Failed to build postgres connection pool")]
|
|
|
|
BuildPool(#[source] BuildError),
|
|
|
|
}
|
|
|
|
|
|
|
|
#[derive(Debug, thiserror::Error)]
|
2023-09-02 18:52:15 +00:00
|
|
|
pub(crate) enum PostgresError {
|
|
|
|
#[error("Error in db pool")]
|
|
|
|
Pool(#[source] PoolError),
|
|
|
|
|
|
|
|
#[error("Error in database")]
|
|
|
|
Diesel(#[source] diesel::result::Error),
|
2023-09-03 17:47:06 +00:00
|
|
|
|
|
|
|
#[error("Error deserializing hex value")]
|
|
|
|
Hex(#[source] hex::FromHexError),
|
|
|
|
|
|
|
|
#[error("Error serializing details")]
|
|
|
|
SerializeDetails(#[source] serde_json::Error),
|
|
|
|
|
|
|
|
#[error("Error deserializing details")]
|
|
|
|
DeserializeDetails(#[source] serde_json::Error),
|
2023-09-03 21:59:41 +00:00
|
|
|
|
|
|
|
#[error("Error serializing upload result")]
|
|
|
|
SerializeUploadResult(#[source] serde_json::Error),
|
|
|
|
|
|
|
|
#[error("Error deserializing upload result")]
|
|
|
|
DeserializeUploadResult(#[source] serde_json::Error),
|
2023-09-05 02:51:27 +00:00
|
|
|
|
|
|
|
#[error("Timed out waiting for postgres")]
|
|
|
|
DbTimeout,
|
2023-09-02 18:52:15 +00:00
|
|
|
}
|
2023-09-02 16:52:55 +00:00
|
|
|
|
2023-09-02 23:30:45 +00:00
|
|
|
impl PostgresError {
|
|
|
|
pub(super) const fn error_code(&self) -> ErrorCode {
|
2023-09-03 22:11:34 +00:00
|
|
|
match self {
|
|
|
|
Self::Pool(_)
|
|
|
|
| Self::Diesel(_)
|
|
|
|
| Self::SerializeDetails(_)
|
|
|
|
| Self::SerializeUploadResult(_)
|
2023-09-05 02:51:27 +00:00
|
|
|
| Self::Hex(_)
|
|
|
|
| Self::DbTimeout => ErrorCode::POSTGRES_ERROR,
|
2023-09-03 22:11:34 +00:00
|
|
|
Self::DeserializeDetails(_) => ErrorCode::EXTRACT_DETAILS,
|
|
|
|
Self::DeserializeUploadResult(_) => ErrorCode::EXTRACT_UPLOAD_RESULT,
|
|
|
|
}
|
2023-09-02 23:30:45 +00:00
|
|
|
}
|
2023-09-05 02:51:27 +00:00
|
|
|
|
|
|
|
pub(super) const fn is_disconnected(&self) -> bool {
|
2023-09-05 03:06:28 +00:00
|
|
|
matches!(
|
|
|
|
self,
|
2023-09-05 02:51:27 +00:00
|
|
|
Self::Pool(
|
|
|
|
PoolError::Closed
|
2023-09-05 03:06:28 +00:00
|
|
|
| PoolError::Backend(
|
|
|
|
diesel_async::pooled_connection::PoolError::ConnectionError(_)
|
|
|
|
),
|
|
|
|
) | Self::Diesel(diesel::result::Error::DatabaseError(
|
2023-09-05 02:51:27 +00:00
|
|
|
diesel::result::DatabaseErrorKind::ClosedConnection,
|
|
|
|
_,
|
2023-09-05 03:06:28 +00:00
|
|
|
))
|
|
|
|
)
|
2023-09-05 02:51:27 +00:00
|
|
|
}
|
2023-09-02 23:30:45 +00:00
|
|
|
}
|
|
|
|
|
2023-09-02 16:52:55 +00:00
|
|
|
impl PostgresRepo {
|
|
|
|
pub(crate) async fn connect(postgres_url: Url) -> Result<Self, ConnectPostgresError> {
|
2023-09-04 02:30:47 +00:00
|
|
|
let (mut client, conn) = tokio_postgres::connect(postgres_url.as_str(), NoTls)
|
|
|
|
.await
|
|
|
|
.map_err(ConnectPostgresError::ConnectForMigration)?;
|
2023-09-02 16:52:55 +00:00
|
|
|
|
2023-09-04 02:30:47 +00:00
|
|
|
let handle = crate::sync::spawn(conn);
|
2023-09-02 16:52:55 +00:00
|
|
|
|
|
|
|
embedded::migrations::runner()
|
|
|
|
.run_async(&mut client)
|
|
|
|
.await
|
|
|
|
.map_err(ConnectPostgresError::Migration)?;
|
|
|
|
|
|
|
|
handle.abort();
|
|
|
|
let _ = handle.await;
|
|
|
|
|
2023-09-05 02:51:27 +00:00
|
|
|
let parallelism = std::thread::available_parallelism()
|
|
|
|
.map(|u| u.into())
|
|
|
|
.unwrap_or(1_usize);
|
|
|
|
|
2023-09-03 17:47:06 +00:00
|
|
|
let (tx, rx) = flume::bounded(10);
|
2023-09-03 01:13:32 +00:00
|
|
|
|
|
|
|
let mut config = ManagerConfig::default();
|
|
|
|
config.custom_setup = build_handler(tx);
|
|
|
|
|
|
|
|
let mgr = AsyncDieselConnectionManager::<AsyncPgConnection>::new_with_config(
|
|
|
|
postgres_url,
|
|
|
|
config,
|
|
|
|
);
|
2023-09-05 02:51:27 +00:00
|
|
|
|
2023-09-03 01:13:32 +00:00
|
|
|
let pool = Pool::builder(mgr)
|
2023-09-05 02:51:27 +00:00
|
|
|
.runtime(deadpool::Runtime::Tokio1)
|
|
|
|
.wait_timeout(Some(Duration::from_secs(1)))
|
|
|
|
.create_timeout(Some(Duration::from_secs(2)))
|
|
|
|
.recycle_timeout(Some(Duration::from_secs(2)))
|
|
|
|
.post_create(Hook::sync_fn(|_, _| {
|
|
|
|
metrics::increment_counter!("pict-rs.postgres.pool.connection.create");
|
|
|
|
Ok(())
|
|
|
|
}))
|
|
|
|
.post_recycle(Hook::sync_fn(|_, _| {
|
|
|
|
metrics::increment_counter!("pict-rs.postgres.pool.connection.recycle");
|
|
|
|
Ok(())
|
|
|
|
}))
|
|
|
|
.max_size(parallelism * 8)
|
2023-09-02 16:52:55 +00:00
|
|
|
.build()
|
|
|
|
.map_err(ConnectPostgresError::BuildPool)?;
|
|
|
|
|
2023-09-03 17:47:06 +00:00
|
|
|
let inner = Arc::new(Inner {
|
2023-09-03 21:59:41 +00:00
|
|
|
health_count: AtomicU64::new(0),
|
2023-09-03 01:13:32 +00:00
|
|
|
pool,
|
2023-09-03 17:47:06 +00:00
|
|
|
queue_notifications: DashMap::new(),
|
2023-09-04 21:19:46 +00:00
|
|
|
upload_notifications: DashMap::new(),
|
2023-09-03 17:47:06 +00:00
|
|
|
});
|
|
|
|
|
2023-09-05 02:51:27 +00:00
|
|
|
let handle = crate::sync::spawn(delegate_notifications(rx, inner.clone(), parallelism * 8));
|
2023-09-04 02:30:47 +00:00
|
|
|
|
|
|
|
let notifications = Arc::new(handle);
|
2023-09-03 17:47:06 +00:00
|
|
|
|
|
|
|
Ok(PostgresRepo {
|
|
|
|
inner,
|
2023-09-03 01:13:32 +00:00
|
|
|
notifications,
|
|
|
|
})
|
2023-09-02 16:52:55 +00:00
|
|
|
}
|
2023-09-04 01:05:29 +00:00
|
|
|
|
|
|
|
async fn get_connection(&self) -> Result<Object<AsyncPgConnection>, PostgresError> {
|
|
|
|
self.inner.get_connection().await
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-09-04 21:19:46 +00:00
|
|
|
struct GetConnectionMetricsGuard {
|
|
|
|
start: Instant,
|
|
|
|
armed: bool,
|
|
|
|
}
|
|
|
|
|
|
|
|
impl GetConnectionMetricsGuard {
|
|
|
|
fn guard() -> Self {
|
|
|
|
GetConnectionMetricsGuard {
|
|
|
|
start: Instant::now(),
|
|
|
|
armed: true,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
fn disarm(mut self) {
|
|
|
|
self.armed = false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
impl Drop for GetConnectionMetricsGuard {
|
|
|
|
fn drop(&mut self) {
|
2023-09-05 02:51:27 +00:00
|
|
|
metrics::increment_counter!("pict-rs.postgres.pool.get", "completed" => (!self.armed).to_string());
|
2023-09-04 21:19:46 +00:00
|
|
|
metrics::histogram!("pict-rs.postgres.pool.get.duration", self.start.elapsed().as_secs_f64(), "completed" => (!self.armed).to_string());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
impl Inner {
|
2023-09-04 21:19:46 +00:00
|
|
|
#[tracing::instrument(level = "TRACE", skip(self))]
|
2023-09-04 01:05:29 +00:00
|
|
|
async fn get_connection(&self) -> Result<Object<AsyncPgConnection>, PostgresError> {
|
2023-09-04 21:19:46 +00:00
|
|
|
let guard = GetConnectionMetricsGuard::guard();
|
2023-09-05 02:51:27 +00:00
|
|
|
|
|
|
|
let obj = self.pool.get().await.map_err(PostgresError::Pool)?;
|
|
|
|
|
2023-09-04 21:19:46 +00:00
|
|
|
guard.disarm();
|
2023-09-05 02:51:27 +00:00
|
|
|
|
|
|
|
Ok(obj)
|
2023-09-04 21:19:46 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
fn interest(self: &Arc<Self>, upload_id: UploadId) -> UploadInterest {
|
|
|
|
let notify = crate::sync::notify();
|
|
|
|
|
|
|
|
self.upload_notifications
|
|
|
|
.insert(upload_id, Arc::downgrade(¬ify));
|
|
|
|
|
|
|
|
UploadInterest {
|
|
|
|
inner: self.clone(),
|
|
|
|
interest: Some(notify),
|
|
|
|
upload_id,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
impl UploadInterest {
|
|
|
|
async fn notified_timeout(&self, timeout: Duration) -> Result<(), tokio::time::error::Elapsed> {
|
2023-09-05 02:58:57 +00:00
|
|
|
self.interest
|
|
|
|
.as_ref()
|
|
|
|
.expect("interest exists")
|
|
|
|
.notified()
|
|
|
|
.with_timeout(timeout)
|
|
|
|
.await
|
2023-09-04 21:19:46 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
impl Drop for UploadInterest {
|
|
|
|
fn drop(&mut self) {
|
|
|
|
if let Some(interest) = self.interest.take() {
|
|
|
|
if Arc::into_inner(interest).is_some() {
|
|
|
|
self.inner.upload_notifications.remove(&self.upload_id);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
impl<'a> JobNotifierState<'a> {
|
|
|
|
fn handle(&mut self, payload: &str) {
|
|
|
|
let Some((job_id, queue_name)) = payload.split_once(' ') else {
|
|
|
|
tracing::warn!("Invalid queue payload {payload}");
|
|
|
|
return;
|
|
|
|
};
|
|
|
|
|
|
|
|
let Ok(job_id) = job_id.parse::<Uuid>().map(JobId) else {
|
|
|
|
tracing::warn!("Invalid job ID {job_id}");
|
|
|
|
return;
|
|
|
|
};
|
|
|
|
|
|
|
|
if !self.jobs.insert(job_id) {
|
|
|
|
// duplicate job
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
self.jobs_ordered.push_back(job_id);
|
|
|
|
|
|
|
|
if self.jobs_ordered.len() > self.capacity {
|
|
|
|
if let Some(job_id) = self.jobs_ordered.pop_front() {
|
|
|
|
self.jobs.remove(&job_id);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
self.inner
|
|
|
|
.queue_notifications
|
|
|
|
.entry(queue_name.to_string())
|
|
|
|
.or_insert_with(crate::sync::notify)
|
|
|
|
.notify_one();
|
|
|
|
|
|
|
|
metrics::increment_counter!("pict-rs.postgres.job-notifier.notified", "queue" => queue_name.to_string());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
impl<'a> UploadNotifierState<'a> {
|
|
|
|
fn handle(&self, payload: &str) {
|
|
|
|
let Ok(upload_id) = payload.parse::<UploadId>() else {
|
|
|
|
tracing::warn!("Invalid upload id {payload}");
|
|
|
|
return;
|
|
|
|
};
|
|
|
|
|
|
|
|
if let Some(notifier) = self
|
|
|
|
.inner
|
|
|
|
.upload_notifications
|
|
|
|
.get(&upload_id)
|
|
|
|
.and_then(|weak| weak.upgrade())
|
|
|
|
{
|
|
|
|
notifier.notify_waiters();
|
|
|
|
metrics::increment_counter!("pict-rs.postgres.upload-notifier.notified");
|
|
|
|
}
|
2023-09-04 01:05:29 +00:00
|
|
|
}
|
2023-09-02 16:52:55 +00:00
|
|
|
}
|
|
|
|
|
2023-09-03 01:13:32 +00:00
|
|
|
type BoxFuture<'a, T> = std::pin::Pin<Box<dyn std::future::Future<Output = T> + Send + 'a>>;
|
|
|
|
type ConfigFn =
|
|
|
|
Box<dyn Fn(&str) -> BoxFuture<'_, ConnectionResult<AsyncPgConnection>> + Send + Sync + 'static>;
|
|
|
|
|
2023-09-05 02:51:27 +00:00
|
|
|
async fn delegate_notifications(
|
|
|
|
receiver: flume::Receiver<Notification>,
|
|
|
|
inner: Arc<Inner>,
|
|
|
|
capacity: usize,
|
|
|
|
) {
|
2023-09-04 21:19:46 +00:00
|
|
|
let mut job_notifier_state = JobNotifierState {
|
|
|
|
inner: &inner,
|
2023-09-05 02:51:27 +00:00
|
|
|
capacity,
|
2023-09-04 21:19:46 +00:00
|
|
|
jobs: BTreeSet::new(),
|
|
|
|
jobs_ordered: VecDeque::new(),
|
|
|
|
};
|
|
|
|
|
|
|
|
let upload_notifier_state = UploadNotifierState { inner: &inner };
|
|
|
|
|
|
|
|
while let Ok(notification) = receiver.recv_async().await {
|
|
|
|
metrics::increment_counter!("pict-rs.postgres.notification");
|
|
|
|
|
|
|
|
match notification.channel() {
|
|
|
|
"queue_status_channel" => {
|
|
|
|
// new job inserted for queue
|
|
|
|
job_notifier_state.handle(notification.payload());
|
|
|
|
}
|
|
|
|
"upload_completion_channel" => {
|
|
|
|
// new upload finished
|
|
|
|
upload_notifier_state.handle(notification.payload());
|
|
|
|
}
|
|
|
|
channel => {
|
|
|
|
tracing::info!(
|
|
|
|
"Unhandled postgres notification: {channel}: {}",
|
|
|
|
notification.payload()
|
|
|
|
);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
tracing::warn!("Notification delegator shutting down");
|
|
|
|
}
|
|
|
|
|
2023-09-03 01:13:32 +00:00
|
|
|
fn build_handler(sender: flume::Sender<Notification>) -> ConfigFn {
|
|
|
|
Box::new(
|
|
|
|
move |config: &str| -> BoxFuture<'_, ConnectionResult<AsyncPgConnection>> {
|
|
|
|
let sender = sender.clone();
|
|
|
|
|
2023-09-04 02:30:47 +00:00
|
|
|
let connect_span = tracing::trace_span!(parent: None, "connect future");
|
2023-09-03 01:13:32 +00:00
|
|
|
|
2023-09-04 02:30:47 +00:00
|
|
|
Box::pin(
|
|
|
|
async move {
|
|
|
|
let (client, conn) =
|
|
|
|
tokio_postgres::connect(config, tokio_postgres::tls::NoTls)
|
|
|
|
.await
|
|
|
|
.map_err(|e| ConnectionError::BadConnection(e.to_string()))?;
|
|
|
|
|
|
|
|
// not very cash money (structured concurrency) of me
|
|
|
|
spawn_db_notification_task(sender, conn);
|
|
|
|
|
|
|
|
AsyncPgConnection::try_from(client).await
|
|
|
|
}
|
|
|
|
.instrument(connect_span),
|
|
|
|
)
|
2023-09-03 01:13:32 +00:00
|
|
|
},
|
|
|
|
)
|
|
|
|
}
|
|
|
|
|
2023-09-04 02:30:47 +00:00
|
|
|
fn spawn_db_notification_task(
|
|
|
|
sender: flume::Sender<Notification>,
|
|
|
|
mut conn: Connection<Socket, NoTlsStream>,
|
|
|
|
) {
|
|
|
|
crate::sync::spawn(async move {
|
|
|
|
while let Some(res) = std::future::poll_fn(|cx| conn.poll_message(cx)).await {
|
|
|
|
match res {
|
|
|
|
Err(e) => {
|
|
|
|
tracing::error!("Database Connection {e:?}");
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
Ok(AsyncMessage::Notice(e)) => {
|
|
|
|
tracing::warn!("Database Notice {e:?}");
|
|
|
|
}
|
|
|
|
Ok(AsyncMessage::Notification(notification)) => {
|
|
|
|
if sender.send_async(notification).await.is_err() {
|
|
|
|
tracing::warn!("Missed notification. Are we shutting down?");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
Ok(_) => {
|
|
|
|
tracing::warn!("Unhandled AsyncMessage!!! Please contact the developer of this application");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
});
|
|
|
|
}
|
|
|
|
|
2023-09-02 23:30:45 +00:00
|
|
|
fn to_primitive(timestamp: time::OffsetDateTime) -> time::PrimitiveDateTime {
|
|
|
|
let timestamp = timestamp.to_offset(time::UtcOffset::UTC);
|
|
|
|
time::PrimitiveDateTime::new(timestamp.date(), timestamp.time())
|
|
|
|
}
|
|
|
|
|
2023-09-02 16:52:55 +00:00
|
|
|
impl BaseRepo for PostgresRepo {}
|
|
|
|
|
2023-09-02 18:52:15 +00:00
|
|
|
#[async_trait::async_trait(?Send)]
|
2023-09-02 16:52:55 +00:00
|
|
|
impl HashRepo for PostgresRepo {
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-02 16:52:55 +00:00
|
|
|
async fn size(&self) -> Result<u64, RepoError> {
|
2023-09-02 18:52:15 +00:00
|
|
|
use schema::hashes::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-02 18:52:15 +00:00
|
|
|
|
|
|
|
let count = hashes
|
|
|
|
.count()
|
|
|
|
.get_result::<i64>(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.hashes.count")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-02 18:52:15 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-02 18:52:15 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(count.try_into().expect("non-negative count"))
|
2023-09-02 16:52:55 +00:00
|
|
|
}
|
2023-09-02 23:30:45 +00:00
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-02 23:30:45 +00:00
|
|
|
async fn bound(&self, input_hash: Hash) -> Result<Option<OrderedHash>, RepoError> {
|
|
|
|
use schema::hashes::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-02 23:30:45 +00:00
|
|
|
|
|
|
|
let timestamp = hashes
|
|
|
|
.select(created_at)
|
|
|
|
.filter(hash.eq(&input_hash))
|
2023-09-05 02:51:27 +00:00
|
|
|
.get_result(&mut conn)
|
|
|
|
.with_metrics("pict-rs.postgres.hashes.bound")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-02 23:30:45 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-02 23:30:45 +00:00
|
|
|
.map(time::PrimitiveDateTime::assume_utc)
|
|
|
|
.optional()
|
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(timestamp.map(|timestamp| OrderedHash {
|
|
|
|
timestamp,
|
|
|
|
hash: input_hash,
|
|
|
|
}))
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-02 23:30:45 +00:00
|
|
|
async fn hash_page_by_date(
|
|
|
|
&self,
|
|
|
|
date: time::OffsetDateTime,
|
|
|
|
limit: usize,
|
|
|
|
) -> Result<HashPage, RepoError> {
|
|
|
|
use schema::hashes::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-02 23:30:45 +00:00
|
|
|
|
|
|
|
let timestamp = to_primitive(date);
|
|
|
|
|
|
|
|
let ordered_hash = hashes
|
|
|
|
.select((created_at, hash))
|
|
|
|
.filter(created_at.lt(timestamp))
|
|
|
|
.order(created_at.desc())
|
2023-09-05 02:51:27 +00:00
|
|
|
.get_result::<(time::PrimitiveDateTime, Hash)>(&mut conn)
|
|
|
|
.with_metrics("pict-rs.postgres.hashes.ordered-hash")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-02 23:30:45 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-02 23:30:45 +00:00
|
|
|
.optional()
|
|
|
|
.map_err(PostgresError::Diesel)?
|
|
|
|
.map(|tup| OrderedHash {
|
|
|
|
timestamp: tup.0.assume_utc(),
|
|
|
|
hash: tup.1,
|
|
|
|
});
|
|
|
|
|
|
|
|
self.hashes_ordered(ordered_hash, limit).await
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-02 23:30:45 +00:00
|
|
|
async fn hashes_ordered(
|
|
|
|
&self,
|
|
|
|
bound: Option<OrderedHash>,
|
|
|
|
limit: usize,
|
|
|
|
) -> Result<HashPage, RepoError> {
|
|
|
|
use schema::hashes::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-02 23:30:45 +00:00
|
|
|
|
|
|
|
let (mut page, prev) = if let Some(OrderedHash {
|
|
|
|
timestamp,
|
|
|
|
hash: bound_hash,
|
|
|
|
}) = bound
|
|
|
|
{
|
|
|
|
let timestamp = to_primitive(timestamp);
|
|
|
|
|
|
|
|
let page = hashes
|
|
|
|
.select(hash)
|
|
|
|
.filter(created_at.lt(timestamp))
|
|
|
|
.or_filter(created_at.eq(timestamp).and(hash.le(&bound_hash)))
|
|
|
|
.order(created_at.desc())
|
|
|
|
.then_order_by(hash.desc())
|
|
|
|
.limit(limit as i64 + 1)
|
2023-09-05 02:51:27 +00:00
|
|
|
.get_results::<Hash>(&mut conn)
|
|
|
|
.with_metrics("pict-rs.postgres.hashes.next-hashes")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-02 23:30:45 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-02 23:30:45 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
let prev = hashes
|
|
|
|
.select(hash)
|
|
|
|
.filter(created_at.gt(timestamp))
|
|
|
|
.or_filter(created_at.eq(timestamp).and(hash.gt(&bound_hash)))
|
|
|
|
.order(created_at)
|
|
|
|
.then_order_by(hash)
|
2023-09-04 02:58:31 +00:00
|
|
|
.limit(limit as i64)
|
|
|
|
.get_results::<Hash>(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.hashes.prev-hashes")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-02 23:30:45 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-04 02:58:31 +00:00
|
|
|
.map_err(PostgresError::Diesel)?
|
|
|
|
.pop();
|
2023-09-02 23:30:45 +00:00
|
|
|
|
|
|
|
(page, prev)
|
|
|
|
} else {
|
|
|
|
let page = hashes
|
|
|
|
.select(hash)
|
|
|
|
.order(created_at.desc())
|
|
|
|
.then_order_by(hash.desc())
|
|
|
|
.limit(limit as i64 + 1)
|
2023-09-05 02:51:27 +00:00
|
|
|
.get_results::<Hash>(&mut conn)
|
|
|
|
.with_metrics("pict-rs.postgres.hashes.first-hashes")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-02 23:30:45 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-02 23:30:45 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
(page, None)
|
|
|
|
};
|
|
|
|
|
|
|
|
let next = if page.len() > limit { page.pop() } else { None };
|
|
|
|
|
|
|
|
Ok(HashPage {
|
|
|
|
limit,
|
|
|
|
prev,
|
|
|
|
next,
|
|
|
|
hashes: page,
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-02 23:30:45 +00:00
|
|
|
async fn create_hash_with_timestamp(
|
|
|
|
&self,
|
|
|
|
input_hash: Hash,
|
|
|
|
input_identifier: &Arc<str>,
|
|
|
|
timestamp: time::OffsetDateTime,
|
|
|
|
) -> Result<Result<(), HashAlreadyExists>, RepoError> {
|
|
|
|
use schema::hashes::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-02 23:30:45 +00:00
|
|
|
|
|
|
|
let timestamp = to_primitive(timestamp);
|
|
|
|
|
2023-09-03 00:13:00 +00:00
|
|
|
let res = diesel::insert_into(hashes)
|
|
|
|
.values((
|
|
|
|
hash.eq(&input_hash),
|
|
|
|
identifier.eq(input_identifier.as_ref()),
|
|
|
|
created_at.eq(×tamp),
|
|
|
|
))
|
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.hashes.create-hash")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-05 02:51:27 +00:00
|
|
|
.await
|
|
|
|
.map_err(|_| PostgresError::DbTimeout)?;
|
2023-09-03 00:13:00 +00:00
|
|
|
|
|
|
|
match res {
|
|
|
|
Ok(_) => Ok(Ok(())),
|
|
|
|
Err(diesel::result::Error::DatabaseError(
|
|
|
|
diesel::result::DatabaseErrorKind::UniqueViolation,
|
|
|
|
_,
|
|
|
|
)) => Ok(Err(HashAlreadyExists)),
|
|
|
|
Err(e) => Err(PostgresError::Diesel(e).into()),
|
|
|
|
}
|
2023-09-02 23:30:45 +00:00
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 00:13:00 +00:00
|
|
|
async fn update_identifier(
|
|
|
|
&self,
|
|
|
|
input_hash: Hash,
|
|
|
|
input_identifier: &Arc<str>,
|
|
|
|
) -> Result<(), RepoError> {
|
|
|
|
use schema::hashes::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 00:13:00 +00:00
|
|
|
|
|
|
|
diesel::update(hashes)
|
|
|
|
.filter(hash.eq(&input_hash))
|
|
|
|
.set(identifier.eq(input_identifier.as_ref()))
|
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.hashes.update-identifier")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 00:13:00 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 00:13:00 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(())
|
2023-09-02 23:30:45 +00:00
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 00:13:00 +00:00
|
|
|
async fn identifier(&self, input_hash: Hash) -> Result<Option<Arc<str>>, RepoError> {
|
|
|
|
use schema::hashes::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 00:13:00 +00:00
|
|
|
|
|
|
|
let opt = hashes
|
|
|
|
.select(identifier)
|
|
|
|
.filter(hash.eq(&input_hash))
|
|
|
|
.get_result::<String>(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.hashes.identifier")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 00:13:00 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 00:13:00 +00:00
|
|
|
.optional()
|
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(opt.map(Arc::from))
|
2023-09-02 23:30:45 +00:00
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-02 23:30:45 +00:00
|
|
|
async fn relate_variant_identifier(
|
|
|
|
&self,
|
2023-09-03 00:13:00 +00:00
|
|
|
input_hash: Hash,
|
|
|
|
input_variant: String,
|
|
|
|
input_identifier: &Arc<str>,
|
2023-09-02 23:30:45 +00:00
|
|
|
) -> Result<Result<(), VariantAlreadyExists>, RepoError> {
|
2023-09-03 00:13:00 +00:00
|
|
|
use schema::variants::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 00:13:00 +00:00
|
|
|
|
|
|
|
let res = diesel::insert_into(variants)
|
|
|
|
.values((
|
|
|
|
hash.eq(&input_hash),
|
|
|
|
variant.eq(&input_variant),
|
|
|
|
identifier.eq(input_identifier.as_ref()),
|
|
|
|
))
|
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.variants.relate-variant-identifier")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-05 02:51:27 +00:00
|
|
|
.await
|
|
|
|
.map_err(|_| PostgresError::DbTimeout)?;
|
2023-09-03 00:13:00 +00:00
|
|
|
|
|
|
|
match res {
|
|
|
|
Ok(_) => Ok(Ok(())),
|
|
|
|
Err(diesel::result::Error::DatabaseError(
|
|
|
|
diesel::result::DatabaseErrorKind::UniqueViolation,
|
|
|
|
_,
|
|
|
|
)) => Ok(Err(VariantAlreadyExists)),
|
|
|
|
Err(e) => Err(PostgresError::Diesel(e).into()),
|
|
|
|
}
|
2023-09-02 23:30:45 +00:00
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-02 23:30:45 +00:00
|
|
|
async fn variant_identifier(
|
|
|
|
&self,
|
2023-09-03 00:13:00 +00:00
|
|
|
input_hash: Hash,
|
|
|
|
input_variant: String,
|
2023-09-02 23:30:45 +00:00
|
|
|
) -> Result<Option<Arc<str>>, RepoError> {
|
2023-09-03 00:13:00 +00:00
|
|
|
use schema::variants::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 00:13:00 +00:00
|
|
|
|
|
|
|
let opt = variants
|
|
|
|
.select(identifier)
|
|
|
|
.filter(hash.eq(&input_hash))
|
|
|
|
.filter(variant.eq(&input_variant))
|
|
|
|
.get_result::<String>(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.variants.identifier")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 00:13:00 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 00:13:00 +00:00
|
|
|
.optional()
|
|
|
|
.map_err(PostgresError::Diesel)?
|
|
|
|
.map(Arc::from);
|
|
|
|
|
|
|
|
Ok(opt)
|
2023-09-02 23:30:45 +00:00
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 00:13:00 +00:00
|
|
|
async fn variants(&self, input_hash: Hash) -> Result<Vec<(String, Arc<str>)>, RepoError> {
|
|
|
|
use schema::variants::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 00:13:00 +00:00
|
|
|
|
|
|
|
let vec = variants
|
|
|
|
.select((variant, identifier))
|
|
|
|
.filter(hash.eq(&input_hash))
|
|
|
|
.get_results::<(String, String)>(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.variants.for-hash")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 00:13:00 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 00:13:00 +00:00
|
|
|
.map_err(PostgresError::Diesel)?
|
|
|
|
.into_iter()
|
|
|
|
.map(|(s, i)| (s, Arc::from(i)))
|
|
|
|
.collect();
|
|
|
|
|
|
|
|
Ok(vec)
|
2023-09-02 23:30:45 +00:00
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 00:13:00 +00:00
|
|
|
async fn remove_variant(
|
|
|
|
&self,
|
|
|
|
input_hash: Hash,
|
|
|
|
input_variant: String,
|
|
|
|
) -> Result<(), RepoError> {
|
|
|
|
use schema::variants::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 00:13:00 +00:00
|
|
|
|
|
|
|
diesel::delete(variants)
|
|
|
|
.filter(hash.eq(&input_hash))
|
|
|
|
.filter(variant.eq(&input_variant))
|
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.variants.remove")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 00:13:00 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 00:13:00 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(())
|
2023-09-02 23:30:45 +00:00
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-02 23:30:45 +00:00
|
|
|
async fn relate_motion_identifier(
|
|
|
|
&self,
|
2023-09-03 00:13:00 +00:00
|
|
|
input_hash: Hash,
|
|
|
|
input_identifier: &Arc<str>,
|
2023-09-02 23:30:45 +00:00
|
|
|
) -> Result<(), RepoError> {
|
2023-09-03 00:13:00 +00:00
|
|
|
use schema::hashes::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 00:13:00 +00:00
|
|
|
|
|
|
|
diesel::update(hashes)
|
|
|
|
.filter(hash.eq(&input_hash))
|
|
|
|
.set(motion_identifier.eq(input_identifier.as_ref()))
|
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.hashes.relate-motion-identifier")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 00:13:00 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 00:13:00 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(())
|
2023-09-02 23:30:45 +00:00
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 00:13:00 +00:00
|
|
|
async fn motion_identifier(&self, input_hash: Hash) -> Result<Option<Arc<str>>, RepoError> {
|
|
|
|
use schema::hashes::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 00:13:00 +00:00
|
|
|
|
|
|
|
let opt = hashes
|
|
|
|
.select(motion_identifier)
|
|
|
|
.filter(hash.eq(&input_hash))
|
|
|
|
.get_result::<Option<String>>(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.hashes.motion-identifier")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 00:13:00 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 00:13:00 +00:00
|
|
|
.optional()
|
|
|
|
.map_err(PostgresError::Diesel)?
|
|
|
|
.flatten()
|
|
|
|
.map(Arc::from);
|
|
|
|
|
|
|
|
Ok(opt)
|
2023-09-02 23:30:45 +00:00
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 00:13:00 +00:00
|
|
|
async fn cleanup_hash(&self, input_hash: Hash) -> Result<(), RepoError> {
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 00:13:00 +00:00
|
|
|
|
|
|
|
conn.transaction(|conn| {
|
|
|
|
Box::pin(async move {
|
2023-09-03 23:21:46 +00:00
|
|
|
diesel::delete(schema::variants::dsl::variants)
|
|
|
|
.filter(schema::variants::dsl::hash.eq(&input_hash))
|
2023-09-03 00:13:00 +00:00
|
|
|
.execute(conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.variants.cleanup")
|
2023-09-03 00:13:00 +00:00
|
|
|
.await?;
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
diesel::delete(schema::hashes::dsl::hashes)
|
|
|
|
.filter(schema::hashes::dsl::hash.eq(&input_hash))
|
2023-09-03 00:13:00 +00:00
|
|
|
.execute(conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.hashes.cleanup")
|
2023-09-03 00:13:00 +00:00
|
|
|
.await
|
|
|
|
})
|
|
|
|
})
|
|
|
|
.await
|
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(())
|
2023-09-02 23:30:45 +00:00
|
|
|
}
|
2023-09-02 16:52:55 +00:00
|
|
|
}
|
|
|
|
|
2023-09-03 17:47:06 +00:00
|
|
|
#[async_trait::async_trait(?Send)]
|
|
|
|
impl AliasRepo for PostgresRepo {
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 17:47:06 +00:00
|
|
|
async fn create_alias(
|
|
|
|
&self,
|
|
|
|
input_alias: &Alias,
|
|
|
|
delete_token: &DeleteToken,
|
|
|
|
input_hash: Hash,
|
|
|
|
) -> Result<Result<(), AliasAlreadyExists>, RepoError> {
|
|
|
|
use schema::aliases::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 17:47:06 +00:00
|
|
|
|
|
|
|
let res = diesel::insert_into(aliases)
|
|
|
|
.values((
|
|
|
|
alias.eq(input_alias),
|
|
|
|
hash.eq(&input_hash),
|
|
|
|
token.eq(delete_token),
|
|
|
|
))
|
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.aliases.create")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-05 02:51:27 +00:00
|
|
|
.await
|
|
|
|
.map_err(|_| PostgresError::DbTimeout)?;
|
2023-09-03 17:47:06 +00:00
|
|
|
|
|
|
|
match res {
|
|
|
|
Ok(_) => Ok(Ok(())),
|
|
|
|
Err(diesel::result::Error::DatabaseError(
|
|
|
|
diesel::result::DatabaseErrorKind::UniqueViolation,
|
|
|
|
_,
|
|
|
|
)) => Ok(Err(AliasAlreadyExists)),
|
|
|
|
Err(e) => Err(PostgresError::Diesel(e).into()),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 17:47:06 +00:00
|
|
|
async fn delete_token(&self, input_alias: &Alias) -> Result<Option<DeleteToken>, RepoError> {
|
|
|
|
use schema::aliases::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 17:47:06 +00:00
|
|
|
|
|
|
|
let opt = aliases
|
|
|
|
.select(token)
|
|
|
|
.filter(alias.eq(input_alias))
|
|
|
|
.get_result(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.aliases.delete-token")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 17:47:06 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 17:47:06 +00:00
|
|
|
.optional()
|
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(opt)
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 17:47:06 +00:00
|
|
|
async fn hash(&self, input_alias: &Alias) -> Result<Option<Hash>, RepoError> {
|
|
|
|
use schema::aliases::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 17:47:06 +00:00
|
|
|
|
|
|
|
let opt = aliases
|
|
|
|
.select(hash)
|
|
|
|
.filter(alias.eq(input_alias))
|
|
|
|
.get_result(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.aliases.hash")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 17:47:06 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 17:47:06 +00:00
|
|
|
.optional()
|
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(opt)
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 17:47:06 +00:00
|
|
|
async fn aliases_for_hash(&self, input_hash: Hash) -> Result<Vec<Alias>, RepoError> {
|
|
|
|
use schema::aliases::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 17:47:06 +00:00
|
|
|
|
|
|
|
let vec = aliases
|
|
|
|
.select(alias)
|
|
|
|
.filter(hash.eq(&input_hash))
|
|
|
|
.get_results(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.aliases.for-hash")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 17:47:06 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 17:47:06 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(vec)
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 17:47:06 +00:00
|
|
|
async fn cleanup_alias(&self, input_alias: &Alias) -> Result<(), RepoError> {
|
|
|
|
use schema::aliases::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 17:47:06 +00:00
|
|
|
|
|
|
|
diesel::delete(aliases)
|
|
|
|
.filter(alias.eq(input_alias))
|
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.aliases.cleanup")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 17:47:06 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 17:47:06 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
#[async_trait::async_trait(?Send)]
|
|
|
|
impl SettingsRepo for PostgresRepo {
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self, input_value))]
|
2023-09-03 17:47:06 +00:00
|
|
|
async fn set(&self, input_key: &'static str, input_value: Arc<[u8]>) -> Result<(), RepoError> {
|
|
|
|
use schema::settings::dsl::*;
|
|
|
|
|
|
|
|
let input_value = hex::encode(input_value);
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 17:47:06 +00:00
|
|
|
|
|
|
|
diesel::insert_into(settings)
|
2023-09-03 23:21:46 +00:00
|
|
|
.values((key.eq(input_key), value.eq(&input_value)))
|
|
|
|
.on_conflict(key)
|
|
|
|
.do_update()
|
|
|
|
.set(value.eq(&input_value))
|
2023-09-03 17:47:06 +00:00
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.settings.set")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 17:47:06 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 17:47:06 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 17:47:06 +00:00
|
|
|
async fn get(&self, input_key: &'static str) -> Result<Option<Arc<[u8]>>, RepoError> {
|
|
|
|
use schema::settings::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 17:47:06 +00:00
|
|
|
|
|
|
|
let opt = settings
|
|
|
|
.select(value)
|
|
|
|
.filter(key.eq(input_key))
|
|
|
|
.get_result::<String>(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.settings.get")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 17:47:06 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 17:47:06 +00:00
|
|
|
.optional()
|
|
|
|
.map_err(PostgresError::Diesel)?
|
|
|
|
.map(hex::decode)
|
|
|
|
.transpose()
|
|
|
|
.map_err(PostgresError::Hex)?
|
|
|
|
.map(Arc::from);
|
|
|
|
|
|
|
|
Ok(opt)
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 17:47:06 +00:00
|
|
|
async fn remove(&self, input_key: &'static str) -> Result<(), RepoError> {
|
|
|
|
use schema::settings::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 17:47:06 +00:00
|
|
|
|
|
|
|
diesel::delete(settings)
|
|
|
|
.filter(key.eq(input_key))
|
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.settings.remove")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 17:47:06 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 17:47:06 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
#[async_trait::async_trait(?Send)]
|
|
|
|
impl DetailsRepo for PostgresRepo {
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self, input_details))]
|
2023-09-03 17:47:06 +00:00
|
|
|
async fn relate_details(
|
|
|
|
&self,
|
|
|
|
input_identifier: &Arc<str>,
|
|
|
|
input_details: &Details,
|
|
|
|
) -> Result<(), RepoError> {
|
|
|
|
use schema::details::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 17:47:06 +00:00
|
|
|
|
|
|
|
let value =
|
|
|
|
serde_json::to_value(&input_details.inner).map_err(PostgresError::SerializeDetails)?;
|
|
|
|
|
|
|
|
diesel::insert_into(details)
|
|
|
|
.values((identifier.eq(input_identifier.as_ref()), json.eq(&value)))
|
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.details.relate")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 17:47:06 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 17:47:06 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 17:47:06 +00:00
|
|
|
async fn details(&self, input_identifier: &Arc<str>) -> Result<Option<Details>, RepoError> {
|
|
|
|
use schema::details::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 17:47:06 +00:00
|
|
|
|
|
|
|
let opt = details
|
|
|
|
.select(json)
|
|
|
|
.filter(identifier.eq(input_identifier.as_ref()))
|
|
|
|
.get_result::<serde_json::Value>(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.details.get")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 17:47:06 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 17:47:06 +00:00
|
|
|
.optional()
|
|
|
|
.map_err(PostgresError::Diesel)?
|
|
|
|
.map(serde_json::from_value)
|
|
|
|
.transpose()
|
|
|
|
.map_err(PostgresError::DeserializeDetails)?
|
|
|
|
.map(|inner| Details { inner });
|
|
|
|
|
|
|
|
Ok(opt)
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 17:47:06 +00:00
|
|
|
async fn cleanup_details(&self, input_identifier: &Arc<str>) -> Result<(), RepoError> {
|
|
|
|
use schema::details::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 17:47:06 +00:00
|
|
|
|
|
|
|
diesel::delete(details)
|
|
|
|
.filter(identifier.eq(input_identifier.as_ref()))
|
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.details.cleanup")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 17:47:06 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 17:47:06 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
#[async_trait::async_trait(?Send)]
|
|
|
|
impl QueueRepo for PostgresRepo {
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self, job_json))]
|
2023-09-03 17:47:06 +00:00
|
|
|
async fn push(
|
|
|
|
&self,
|
|
|
|
queue_name: &'static str,
|
|
|
|
job_json: serde_json::Value,
|
|
|
|
) -> Result<JobId, RepoError> {
|
2023-09-04 21:19:46 +00:00
|
|
|
let guard = PushMetricsGuard::guard(queue_name);
|
|
|
|
|
2023-09-03 17:47:06 +00:00
|
|
|
use schema::job_queue::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 17:47:06 +00:00
|
|
|
|
|
|
|
let job_id = diesel::insert_into(job_queue)
|
|
|
|
.values((queue.eq(queue_name), job.eq(job_json)))
|
|
|
|
.returning(id)
|
|
|
|
.get_result::<Uuid>(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.queue.push")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 17:47:06 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 17:47:06 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
2023-09-04 21:19:46 +00:00
|
|
|
guard.disarm();
|
|
|
|
|
2023-09-03 17:47:06 +00:00
|
|
|
Ok(JobId(job_id))
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 17:47:06 +00:00
|
|
|
async fn pop(
|
|
|
|
&self,
|
|
|
|
queue_name: &'static str,
|
|
|
|
worker_id: Uuid,
|
|
|
|
) -> Result<(JobId, serde_json::Value), RepoError> {
|
2023-09-04 21:19:46 +00:00
|
|
|
let guard = PopMetricsGuard::guard(queue_name);
|
|
|
|
|
2023-09-03 17:47:06 +00:00
|
|
|
use schema::job_queue::dsl::*;
|
|
|
|
|
|
|
|
loop {
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 17:47:06 +00:00
|
|
|
|
|
|
|
let notifier: Arc<Notify> = self
|
|
|
|
.inner
|
|
|
|
.queue_notifications
|
|
|
|
.entry(String::from(queue_name))
|
2023-09-04 02:30:47 +00:00
|
|
|
.or_insert_with(crate::sync::notify)
|
2023-09-03 17:47:06 +00:00
|
|
|
.clone();
|
|
|
|
|
|
|
|
diesel::sql_query("LISTEN queue_status_channel;")
|
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.queue.listen")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 17:47:06 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 17:47:06 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
let timestamp = to_primitive(time::OffsetDateTime::now_utc());
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let count = diesel::update(job_queue)
|
2023-09-03 17:47:06 +00:00
|
|
|
.filter(heartbeat.le(timestamp.saturating_sub(time::Duration::minutes(2))))
|
|
|
|
.set((
|
|
|
|
heartbeat.eq(Option::<time::PrimitiveDateTime>::None),
|
|
|
|
status.eq(JobStatus::New),
|
|
|
|
))
|
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.queue.requeue")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 17:47:06 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 17:47:06 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
if count > 0 {
|
|
|
|
tracing::info!("Reset {count} jobs");
|
|
|
|
}
|
2023-09-03 17:47:06 +00:00
|
|
|
|
2023-09-08 00:20:41 +00:00
|
|
|
let queue_alias = diesel::alias!(schema::job_queue as queue_alias);
|
|
|
|
|
|
|
|
let id_query = queue_alias
|
|
|
|
.select(queue_alias.field(id))
|
|
|
|
.filter(
|
|
|
|
queue_alias
|
|
|
|
.field(status)
|
|
|
|
.eq(JobStatus::New)
|
|
|
|
.and(queue_alias.field(queue).eq(queue_name)),
|
|
|
|
)
|
|
|
|
.order(queue_alias.field(queue_time))
|
|
|
|
.for_update()
|
|
|
|
.skip_locked()
|
|
|
|
.single_value();
|
|
|
|
|
|
|
|
let opt = diesel::update(job_queue)
|
|
|
|
.filter(id.nullable().eq(id_query))
|
|
|
|
.filter(status.eq(JobStatus::New))
|
|
|
|
.set((
|
|
|
|
heartbeat.eq(timestamp),
|
|
|
|
status.eq(JobStatus::Running),
|
|
|
|
worker.eq(worker_id),
|
|
|
|
))
|
|
|
|
.returning((id, job))
|
|
|
|
.get_result(&mut conn)
|
|
|
|
.with_metrics("pict-rs.postgres.queue.claim")
|
|
|
|
.with_timeout(Duration::from_secs(5))
|
|
|
|
.await
|
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
|
|
|
.optional()
|
|
|
|
.map_err(PostgresError::Diesel)?;
|
2023-09-04 01:05:29 +00:00
|
|
|
|
|
|
|
if let Some((job_id, job_json)) = opt {
|
2023-09-04 21:19:46 +00:00
|
|
|
guard.disarm();
|
2023-09-03 17:47:06 +00:00
|
|
|
return Ok((JobId(job_id), job_json));
|
|
|
|
}
|
|
|
|
|
|
|
|
drop(conn);
|
2023-09-05 02:58:57 +00:00
|
|
|
if notifier
|
|
|
|
.notified()
|
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-04 01:05:29 +00:00
|
|
|
.await
|
|
|
|
.is_ok()
|
|
|
|
{
|
|
|
|
tracing::debug!("Notified");
|
|
|
|
} else {
|
|
|
|
tracing::debug!("Timed out");
|
|
|
|
}
|
2023-09-03 17:47:06 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 17:47:06 +00:00
|
|
|
async fn heartbeat(
|
|
|
|
&self,
|
|
|
|
queue_name: &'static str,
|
|
|
|
worker_id: Uuid,
|
|
|
|
job_id: JobId,
|
|
|
|
) -> Result<(), RepoError> {
|
|
|
|
use schema::job_queue::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 17:47:06 +00:00
|
|
|
|
|
|
|
let timestamp = to_primitive(time::OffsetDateTime::now_utc());
|
|
|
|
|
|
|
|
diesel::update(job_queue)
|
|
|
|
.filter(
|
|
|
|
id.eq(job_id.0)
|
|
|
|
.and(queue.eq(queue_name))
|
|
|
|
.and(worker.eq(worker_id)),
|
|
|
|
)
|
|
|
|
.set(heartbeat.eq(timestamp))
|
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.queue.heartbeat")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 17:47:06 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 17:47:06 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 17:47:06 +00:00
|
|
|
async fn complete_job(
|
|
|
|
&self,
|
|
|
|
queue_name: &'static str,
|
|
|
|
worker_id: Uuid,
|
|
|
|
job_id: JobId,
|
|
|
|
) -> Result<(), RepoError> {
|
|
|
|
use schema::job_queue::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 17:47:06 +00:00
|
|
|
|
|
|
|
diesel::delete(job_queue)
|
|
|
|
.filter(
|
|
|
|
id.eq(job_id.0)
|
|
|
|
.and(queue.eq(queue_name))
|
|
|
|
.and(worker.eq(worker_id)),
|
|
|
|
)
|
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.queue.complete")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 17:47:06 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 17:47:06 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
#[async_trait::async_trait(?Send)]
|
|
|
|
impl StoreMigrationRepo for PostgresRepo {
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 17:47:06 +00:00
|
|
|
async fn is_continuing_migration(&self) -> Result<bool, RepoError> {
|
|
|
|
use schema::store_migrations::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 17:47:06 +00:00
|
|
|
|
|
|
|
let count = store_migrations
|
|
|
|
.count()
|
|
|
|
.get_result::<i64>(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.store-migration.count")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 17:47:06 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 17:47:06 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(count > 0)
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 17:47:06 +00:00
|
|
|
async fn mark_migrated(
|
|
|
|
&self,
|
|
|
|
input_old_identifier: &Arc<str>,
|
|
|
|
input_new_identifier: &Arc<str>,
|
|
|
|
) -> Result<(), RepoError> {
|
|
|
|
use schema::store_migrations::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 17:47:06 +00:00
|
|
|
|
|
|
|
diesel::insert_into(store_migrations)
|
|
|
|
.values((
|
|
|
|
old_identifier.eq(input_old_identifier.as_ref()),
|
|
|
|
new_identifier.eq(input_new_identifier.as_ref()),
|
|
|
|
))
|
2023-09-07 01:49:00 +00:00
|
|
|
.on_conflict(old_identifier)
|
2023-09-03 17:47:06 +00:00
|
|
|
.do_nothing()
|
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.store-migration.mark-migrated")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 17:47:06 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 17:47:06 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 17:47:06 +00:00
|
|
|
async fn is_migrated(&self, input_old_identifier: &Arc<str>) -> Result<bool, RepoError> {
|
|
|
|
use schema::store_migrations::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 17:47:06 +00:00
|
|
|
|
|
|
|
let b = diesel::select(diesel::dsl::exists(
|
|
|
|
store_migrations.filter(old_identifier.eq(input_old_identifier.as_ref())),
|
|
|
|
))
|
|
|
|
.get_result(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.store-migration.is-migrated")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 17:47:06 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 17:47:06 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(b)
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 17:47:06 +00:00
|
|
|
async fn clear(&self) -> Result<(), RepoError> {
|
|
|
|
use schema::store_migrations::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 17:47:06 +00:00
|
|
|
|
|
|
|
diesel::delete(store_migrations)
|
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.store-migration.clear")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(20))
|
2023-09-03 17:47:06 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 17:47:06 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-09-03 21:59:41 +00:00
|
|
|
#[async_trait::async_trait(?Send)]
|
|
|
|
impl ProxyRepo for PostgresRepo {
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 21:59:41 +00:00
|
|
|
async fn relate_url(&self, input_url: Url, input_alias: Alias) -> Result<(), RepoError> {
|
|
|
|
use schema::proxies::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 21:59:41 +00:00
|
|
|
|
|
|
|
diesel::insert_into(proxies)
|
|
|
|
.values((url.eq(input_url.as_str()), alias.eq(&input_alias)))
|
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.proxy.relate-url")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 21:59:41 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 21:59:41 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 21:59:41 +00:00
|
|
|
async fn related(&self, input_url: Url) -> Result<Option<Alias>, RepoError> {
|
|
|
|
use schema::proxies::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 21:59:41 +00:00
|
|
|
|
|
|
|
let opt = proxies
|
|
|
|
.select(alias)
|
|
|
|
.filter(url.eq(input_url.as_str()))
|
|
|
|
.get_result(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.proxy.related")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 21:59:41 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 21:59:41 +00:00
|
|
|
.optional()
|
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(opt)
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 21:59:41 +00:00
|
|
|
async fn remove_relation(&self, input_alias: Alias) -> Result<(), RepoError> {
|
|
|
|
use schema::proxies::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 21:59:41 +00:00
|
|
|
|
|
|
|
diesel::delete(proxies)
|
|
|
|
.filter(alias.eq(&input_alias))
|
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.proxy.remove-relation")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 21:59:41 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 21:59:41 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
#[async_trait::async_trait(?Send)]
|
|
|
|
impl AliasAccessRepo for PostgresRepo {
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 21:59:41 +00:00
|
|
|
async fn set_accessed_alias(
|
|
|
|
&self,
|
|
|
|
input_alias: Alias,
|
|
|
|
timestamp: time::OffsetDateTime,
|
|
|
|
) -> Result<(), RepoError> {
|
|
|
|
use schema::proxies::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 21:59:41 +00:00
|
|
|
|
|
|
|
let timestamp = to_primitive(timestamp);
|
|
|
|
|
|
|
|
diesel::update(proxies)
|
|
|
|
.filter(alias.eq(&input_alias))
|
|
|
|
.set(accessed.eq(timestamp))
|
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.alias-access.set-accessed")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 21:59:41 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 21:59:41 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 21:59:41 +00:00
|
|
|
async fn alias_accessed_at(
|
|
|
|
&self,
|
|
|
|
input_alias: Alias,
|
|
|
|
) -> Result<Option<time::OffsetDateTime>, RepoError> {
|
|
|
|
use schema::proxies::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 21:59:41 +00:00
|
|
|
|
|
|
|
let opt = proxies
|
|
|
|
.select(accessed)
|
|
|
|
.filter(alias.eq(&input_alias))
|
|
|
|
.get_result::<time::PrimitiveDateTime>(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.alias-access.accessed-at")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 21:59:41 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 21:59:41 +00:00
|
|
|
.optional()
|
|
|
|
.map_err(PostgresError::Diesel)?
|
|
|
|
.map(time::PrimitiveDateTime::assume_utc);
|
|
|
|
|
|
|
|
Ok(opt)
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 21:59:41 +00:00
|
|
|
async fn older_aliases(
|
|
|
|
&self,
|
|
|
|
timestamp: time::OffsetDateTime,
|
|
|
|
) -> Result<LocalBoxStream<'static, Result<Alias, RepoError>>, RepoError> {
|
2023-09-11 00:08:01 +00:00
|
|
|
Ok(Box::pin(page_stream(
|
|
|
|
self.inner.clone(),
|
|
|
|
to_primitive(timestamp),
|
|
|
|
|inner, older_than| async move {
|
|
|
|
use schema::proxies::dsl::*;
|
|
|
|
|
|
|
|
let mut conn = inner.get_connection().await?;
|
|
|
|
|
|
|
|
let vec = proxies
|
|
|
|
.select((accessed, alias))
|
|
|
|
.filter(accessed.lt(older_than))
|
|
|
|
.order(accessed.desc())
|
|
|
|
.limit(100)
|
|
|
|
.get_results(&mut conn)
|
|
|
|
.with_metrics("pict-rs.postgres.alias-access.older-aliases")
|
|
|
|
.with_timeout(Duration::from_secs(5))
|
|
|
|
.await
|
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(vec)
|
|
|
|
},
|
|
|
|
)))
|
2023-09-03 21:59:41 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
async fn remove_alias_access(&self, _: Alias) -> Result<(), RepoError> {
|
|
|
|
// Noop - handled by ProxyRepo::remove_relation
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
#[async_trait::async_trait(?Send)]
|
|
|
|
impl VariantAccessRepo for PostgresRepo {
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 21:59:41 +00:00
|
|
|
async fn set_accessed_variant(
|
|
|
|
&self,
|
|
|
|
input_hash: Hash,
|
|
|
|
input_variant: String,
|
|
|
|
input_accessed: time::OffsetDateTime,
|
|
|
|
) -> Result<(), RepoError> {
|
|
|
|
use schema::variants::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 21:59:41 +00:00
|
|
|
|
|
|
|
let timestamp = to_primitive(input_accessed);
|
|
|
|
|
|
|
|
diesel::update(variants)
|
|
|
|
.filter(hash.eq(&input_hash).and(variant.eq(&input_variant)))
|
|
|
|
.set(accessed.eq(timestamp))
|
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.variant-access.set-accessed")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 21:59:41 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 21:59:41 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 21:59:41 +00:00
|
|
|
async fn variant_accessed_at(
|
|
|
|
&self,
|
|
|
|
input_hash: Hash,
|
|
|
|
input_variant: String,
|
|
|
|
) -> Result<Option<time::OffsetDateTime>, RepoError> {
|
|
|
|
use schema::variants::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 21:59:41 +00:00
|
|
|
|
|
|
|
let opt = variants
|
|
|
|
.select(accessed)
|
|
|
|
.filter(hash.eq(&input_hash).and(variant.eq(&input_variant)))
|
|
|
|
.get_result(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.variant-access.accessed-at")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 21:59:41 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 21:59:41 +00:00
|
|
|
.optional()
|
|
|
|
.map_err(PostgresError::Diesel)?
|
|
|
|
.map(time::PrimitiveDateTime::assume_utc);
|
|
|
|
|
|
|
|
Ok(opt)
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 21:59:41 +00:00
|
|
|
async fn older_variants(
|
|
|
|
&self,
|
|
|
|
timestamp: time::OffsetDateTime,
|
|
|
|
) -> Result<LocalBoxStream<'static, Result<(Hash, String), RepoError>>, RepoError> {
|
2023-09-11 00:08:01 +00:00
|
|
|
Ok(Box::pin(page_stream(
|
|
|
|
self.inner.clone(),
|
|
|
|
to_primitive(timestamp),
|
|
|
|
|inner, older_than| async move {
|
|
|
|
use schema::variants::dsl::*;
|
|
|
|
|
|
|
|
let mut conn = inner.get_connection().await?;
|
|
|
|
|
|
|
|
let vec = variants
|
|
|
|
.select((accessed, (hash, variant)))
|
|
|
|
.filter(accessed.lt(older_than))
|
|
|
|
.order(accessed.desc())
|
|
|
|
.limit(100)
|
|
|
|
.get_results(&mut conn)
|
|
|
|
.with_metrics("pict-rs.postgres.variant-access.older-variants")
|
|
|
|
.with_timeout(Duration::from_secs(5))
|
|
|
|
.await
|
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(vec)
|
|
|
|
},
|
|
|
|
)))
|
2023-09-03 21:59:41 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
async fn remove_variant_access(&self, _: Hash, _: String) -> Result<(), RepoError> {
|
|
|
|
// Noop - handled by HashRepo::remove_variant
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
#[derive(serde::Deserialize, serde::Serialize)]
|
|
|
|
enum InnerUploadResult {
|
|
|
|
Success {
|
|
|
|
alias: Serde<Alias>,
|
|
|
|
token: Serde<DeleteToken>,
|
|
|
|
},
|
|
|
|
Failure {
|
|
|
|
message: String,
|
|
|
|
code: OwnedErrorCode,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
impl From<UploadResult> for InnerUploadResult {
|
|
|
|
fn from(u: UploadResult) -> Self {
|
|
|
|
match u {
|
|
|
|
UploadResult::Success { alias, token } => InnerUploadResult::Success {
|
|
|
|
alias: Serde::new(alias),
|
|
|
|
token: Serde::new(token),
|
|
|
|
},
|
|
|
|
UploadResult::Failure { message, code } => InnerUploadResult::Failure { message, code },
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
impl From<InnerUploadResult> for UploadResult {
|
|
|
|
fn from(i: InnerUploadResult) -> Self {
|
|
|
|
match i {
|
|
|
|
InnerUploadResult::Success { alias, token } => UploadResult::Success {
|
|
|
|
alias: Serde::into_inner(alias),
|
|
|
|
token: Serde::into_inner(token),
|
|
|
|
},
|
|
|
|
InnerUploadResult::Failure { message, code } => UploadResult::Failure { message, code },
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
#[async_trait::async_trait(?Send)]
|
|
|
|
impl UploadRepo for PostgresRepo {
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 21:59:41 +00:00
|
|
|
async fn create_upload(&self) -> Result<UploadId, RepoError> {
|
|
|
|
use schema::uploads::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 21:59:41 +00:00
|
|
|
|
|
|
|
let uuid = diesel::insert_into(uploads)
|
|
|
|
.default_values()
|
|
|
|
.returning(id)
|
|
|
|
.get_result(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.uploads.create")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 21:59:41 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 21:59:41 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(UploadId { id: uuid })
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 21:59:41 +00:00
|
|
|
async fn wait(&self, upload_id: UploadId) -> Result<UploadResult, RepoError> {
|
2023-09-04 21:19:46 +00:00
|
|
|
let guard = WaitMetricsGuard::guard();
|
2023-09-03 21:59:41 +00:00
|
|
|
use schema::uploads::dsl::*;
|
|
|
|
|
2023-09-04 21:19:46 +00:00
|
|
|
let interest = self.inner.interest(upload_id);
|
|
|
|
|
2023-09-03 21:59:41 +00:00
|
|
|
loop {
|
2023-09-04 21:19:46 +00:00
|
|
|
let interest_future = interest.notified_timeout(Duration::from_secs(5));
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 21:59:41 +00:00
|
|
|
|
|
|
|
diesel::sql_query("LISTEN upload_completion_channel;")
|
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.uploads.listen")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 21:59:41 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 21:59:41 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
2023-09-04 02:30:47 +00:00
|
|
|
let nested_opt = uploads
|
2023-09-03 21:59:41 +00:00
|
|
|
.select(result)
|
|
|
|
.filter(id.eq(upload_id.id))
|
|
|
|
.get_result(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.uploads.wait")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 21:59:41 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 21:59:41 +00:00
|
|
|
.optional()
|
2023-09-04 02:30:47 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
2023-09-03 21:59:41 +00:00
|
|
|
|
2023-09-04 02:30:47 +00:00
|
|
|
match nested_opt {
|
|
|
|
Some(opt) => {
|
|
|
|
if let Some(upload_result) = opt {
|
|
|
|
let upload_result: InnerUploadResult =
|
|
|
|
serde_json::from_value(upload_result)
|
|
|
|
.map_err(PostgresError::DeserializeUploadResult)?;
|
2023-09-03 21:59:41 +00:00
|
|
|
|
2023-09-04 21:19:46 +00:00
|
|
|
guard.disarm();
|
2023-09-04 02:30:47 +00:00
|
|
|
return Ok(upload_result.into());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
None => {
|
|
|
|
return Err(RepoError::AlreadyClaimed);
|
|
|
|
}
|
2023-09-03 21:59:41 +00:00
|
|
|
}
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
drop(conn);
|
|
|
|
|
2023-09-04 21:19:46 +00:00
|
|
|
if interest_future.await.is_ok() {
|
2023-09-04 01:05:29 +00:00
|
|
|
tracing::debug!("Notified");
|
|
|
|
} else {
|
|
|
|
tracing::debug!("Timed out");
|
|
|
|
}
|
2023-09-03 21:59:41 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 21:59:41 +00:00
|
|
|
async fn claim(&self, upload_id: UploadId) -> Result<(), RepoError> {
|
|
|
|
use schema::uploads::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 21:59:41 +00:00
|
|
|
|
|
|
|
diesel::delete(uploads)
|
|
|
|
.filter(id.eq(upload_id.id))
|
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.uploads.claim")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 21:59:41 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 21:59:41 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 21:59:41 +00:00
|
|
|
async fn complete_upload(
|
|
|
|
&self,
|
|
|
|
upload_id: UploadId,
|
|
|
|
upload_result: UploadResult,
|
|
|
|
) -> Result<(), RepoError> {
|
|
|
|
use schema::uploads::dsl::*;
|
|
|
|
|
2023-09-04 01:05:29 +00:00
|
|
|
let mut conn = self.get_connection().await?;
|
2023-09-03 21:59:41 +00:00
|
|
|
|
|
|
|
let upload_result: InnerUploadResult = upload_result.into();
|
|
|
|
let upload_result =
|
|
|
|
serde_json::to_value(&upload_result).map_err(PostgresError::SerializeUploadResult)?;
|
|
|
|
|
|
|
|
diesel::update(uploads)
|
|
|
|
.filter(id.eq(upload_id.id))
|
|
|
|
.set(result.eq(upload_result))
|
|
|
|
.execute(&mut conn)
|
2023-09-05 02:51:27 +00:00
|
|
|
.with_metrics("pict-rs.postgres.uploads.complete")
|
2023-09-05 03:02:59 +00:00
|
|
|
.with_timeout(Duration::from_secs(5))
|
2023-09-03 21:59:41 +00:00
|
|
|
.await
|
2023-09-05 02:51:27 +00:00
|
|
|
.map_err(|_| PostgresError::DbTimeout)?
|
2023-09-03 21:59:41 +00:00
|
|
|
.map_err(PostgresError::Diesel)?;
|
|
|
|
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
#[async_trait::async_trait(?Send)]
|
|
|
|
impl FullRepo for PostgresRepo {
|
2023-09-03 23:21:46 +00:00
|
|
|
#[tracing::instrument(level = "DEBUG", skip(self))]
|
2023-09-03 21:59:41 +00:00
|
|
|
async fn health_check(&self) -> Result<(), RepoError> {
|
|
|
|
let next = self.inner.health_count.fetch_add(1, Ordering::Relaxed);
|
|
|
|
|
|
|
|
self.set("health-value", Arc::from(next.to_be_bytes()))
|
|
|
|
.await?;
|
|
|
|
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-09-11 00:08:01 +00:00
|
|
|
fn page_stream<I, F, Fut>(
|
2023-09-03 21:59:41 +00:00
|
|
|
inner: Arc<Inner>,
|
2023-09-11 00:08:01 +00:00
|
|
|
mut older_than: time::PrimitiveDateTime,
|
|
|
|
next: F,
|
|
|
|
) -> impl Stream<Item = Result<I, RepoError>>
|
2023-09-03 21:59:41 +00:00
|
|
|
where
|
2023-09-11 00:08:01 +00:00
|
|
|
F: Fn(Arc<Inner>, time::PrimitiveDateTime) -> Fut,
|
|
|
|
Fut: std::future::Future<Output = Result<Vec<(time::PrimitiveDateTime, I)>, RepoError>>
|
|
|
|
+ 'static,
|
|
|
|
I: 'static,
|
2023-09-03 21:59:41 +00:00
|
|
|
{
|
2023-09-11 00:08:01 +00:00
|
|
|
streem::try_from_fn(|yielder| async move {
|
2023-09-03 21:59:41 +00:00
|
|
|
loop {
|
2023-09-11 00:08:01 +00:00
|
|
|
let mut page = (next)(inner.clone(), older_than).await?;
|
2023-09-03 21:59:41 +00:00
|
|
|
|
2023-09-11 00:08:01 +00:00
|
|
|
if let Some((last_time, last_item)) = page.pop() {
|
|
|
|
for (_, item) in page {
|
|
|
|
yielder.yield_ok(item).await;
|
|
|
|
}
|
2023-09-03 21:59:41 +00:00
|
|
|
|
2023-09-11 00:08:01 +00:00
|
|
|
yielder.yield_ok(last_item).await;
|
2023-09-03 21:59:41 +00:00
|
|
|
|
2023-09-11 00:08:01 +00:00
|
|
|
older_than = last_time;
|
2023-09-03 21:59:41 +00:00
|
|
|
} else {
|
2023-09-11 00:08:01 +00:00
|
|
|
break;
|
2023-09-03 21:59:41 +00:00
|
|
|
}
|
|
|
|
}
|
2023-09-11 00:08:01 +00:00
|
|
|
|
|
|
|
Ok(())
|
|
|
|
})
|
2023-09-03 21:59:41 +00:00
|
|
|
}
|
|
|
|
|
2023-09-02 16:52:55 +00:00
|
|
|
impl std::fmt::Debug for PostgresRepo {
|
|
|
|
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
|
|
|
f.debug_struct("PostgresRepo")
|
|
|
|
.field("pool", &"pool")
|
|
|
|
.finish()
|
|
|
|
}
|
|
|
|
}
|