mirror of
https://git.asonix.dog/asonix/pict-rs
synced 2024-11-20 11:21:14 +00:00
Finish media proxy implementation
This commit is contained in:
parent
4cca7d0f86
commit
aac1bb7bc4
3 changed files with 128 additions and 32 deletions
120
src/lib.rs
120
src/lib.rs
|
@ -40,6 +40,7 @@ use futures_util::{
|
|||
use metrics_exporter_prometheus::PrometheusBuilder;
|
||||
use middleware::Metrics;
|
||||
use once_cell::sync::Lazy;
|
||||
use repo::AliasAccessRepo;
|
||||
use reqwest_middleware::{ClientBuilder, ClientWithMiddleware};
|
||||
use reqwest_tracing::TracingMiddleware;
|
||||
use rusty_s3::UrlStyle;
|
||||
|
@ -482,6 +483,24 @@ struct UrlQuery {
|
|||
backgrounded: bool,
|
||||
}
|
||||
|
||||
async fn ingest_inline<R: FullRepo, S: Store + 'static>(
|
||||
stream: impl Stream<Item = Result<web::Bytes, Error>> + Unpin + 'static,
|
||||
repo: &R,
|
||||
store: &S,
|
||||
config: &Configuration,
|
||||
) -> Result<(Alias, DeleteToken, Details), Error> {
|
||||
let mut session = ingest::ingest(repo, store, stream, None, &config.media).await?;
|
||||
|
||||
let alias = session.alias().expect("alias should exist").to_owned();
|
||||
let delete_token = session.delete_token().await?;
|
||||
|
||||
let details = ensure_details(repo, store, config, &alias).await?;
|
||||
|
||||
session.disarm();
|
||||
|
||||
Ok((alias, delete_token, details))
|
||||
}
|
||||
|
||||
/// download an image from a URL
|
||||
#[tracing::instrument(name = "Downloading file", skip(client, repo, store, config))]
|
||||
async fn download<R: FullRepo + 'static, S: Store + 'static>(
|
||||
|
@ -491,11 +510,25 @@ async fn download<R: FullRepo + 'static, S: Store + 'static>(
|
|||
config: web::Data<Configuration>,
|
||||
query: web::Query<UrlQuery>,
|
||||
) -> Result<HttpResponse, Error> {
|
||||
let stream = download_stream(client, &query.url, &config).await?;
|
||||
|
||||
if query.backgrounded {
|
||||
do_download_backgrounded(stream, repo, store).await
|
||||
} else {
|
||||
do_download_inline(stream, repo, store, config).await
|
||||
}
|
||||
}
|
||||
|
||||
async fn download_stream(
|
||||
client: web::Data<ClientWithMiddleware>,
|
||||
url: &str,
|
||||
config: &Configuration,
|
||||
) -> Result<impl Stream<Item = Result<web::Bytes, Error>> + Unpin + 'static, Error> {
|
||||
if config.server.read_only {
|
||||
return Err(UploadError::ReadOnly.into());
|
||||
}
|
||||
|
||||
let res = client.get(&query.url).send().await?;
|
||||
let res = client.get(url).send().await?;
|
||||
|
||||
if !res.status().is_success() {
|
||||
return Err(UploadError::Download(res.status()).into());
|
||||
|
@ -506,15 +539,11 @@ async fn download<R: FullRepo + 'static, S: Store + 'static>(
|
|||
.map_err(Error::from)
|
||||
.limit((config.media.max_file_size * MEGABYTES) as u64);
|
||||
|
||||
if query.backgrounded {
|
||||
do_download_backgrounded(stream, repo, store).await
|
||||
} else {
|
||||
do_download_inline(stream, repo, store, config).await
|
||||
}
|
||||
Ok(stream)
|
||||
}
|
||||
|
||||
#[tracing::instrument(name = "Downloading file inline", skip(stream, repo, store, config))]
|
||||
async fn do_download_inline<R: FullRepo + 'static, S: Store + 'static>(
|
||||
async fn do_download_inline<R: FullRepo, S: Store + 'static>(
|
||||
stream: impl Stream<Item = Result<web::Bytes, Error>> + Unpin + 'static,
|
||||
repo: web::Data<R>,
|
||||
store: web::Data<S>,
|
||||
|
@ -522,14 +551,7 @@ async fn do_download_inline<R: FullRepo + 'static, S: Store + 'static>(
|
|||
) -> Result<HttpResponse, Error> {
|
||||
metrics::increment_counter!("pict-rs.files", "download" => "inline");
|
||||
|
||||
let mut session = ingest::ingest(&repo, &store, stream, None, &config.media).await?;
|
||||
|
||||
let alias = session.alias().expect("alias should exist").to_owned();
|
||||
let delete_token = session.delete_token().await?;
|
||||
|
||||
let details = ensure_details(&repo, &store, &config, &alias).await?;
|
||||
|
||||
session.disarm();
|
||||
let (alias, delete_token, details) = ingest_inline(stream, &repo, &store, &config).await?;
|
||||
|
||||
Ok(HttpResponse::Created().json(&serde_json::json!({
|
||||
"msg": "ok",
|
||||
|
@ -660,8 +682,7 @@ async fn process_details<R: FullRepo, S: Store>(
|
|||
let thumbnail_string = thumbnail_path.to_string_lossy().to_string();
|
||||
|
||||
if !config.server.read_only {
|
||||
repo.accessed(hash.clone(), thumbnail_string.clone())
|
||||
.await?;
|
||||
VariantAccessRepo::accessed(&repo, hash.clone(), thumbnail_string.clone()).await?;
|
||||
}
|
||||
|
||||
let identifier = repo
|
||||
|
@ -697,7 +718,7 @@ async fn not_found_hash<R: FullRepo>(repo: &R) -> Result<Option<(Alias, R::Bytes
|
|||
/// Process files
|
||||
#[tracing::instrument(
|
||||
name = "Serving processed image",
|
||||
skip(repo, store, config, process_map)
|
||||
skip(repo, store, client, config, process_map)
|
||||
)]
|
||||
async fn process<R: FullRepo, S: Store + 'static>(
|
||||
range: Option<web::Header<Range>>,
|
||||
|
@ -705,6 +726,7 @@ async fn process<R: FullRepo, S: Store + 'static>(
|
|||
ext: web::Path<String>,
|
||||
repo: web::Data<R>,
|
||||
store: web::Data<S>,
|
||||
client: web::Data<ClientWithMiddleware>,
|
||||
config: web::Data<Configuration>,
|
||||
process_map: web::Data<ProcessMap>,
|
||||
) -> Result<HttpResponse, Error> {
|
||||
|
@ -713,11 +735,25 @@ async fn process<R: FullRepo, S: Store + 'static>(
|
|||
Serde::into_inner(alias)
|
||||
}
|
||||
ProcessSource::Proxy { proxy } => {
|
||||
if let Some(alias) = repo.related(proxy).await? {
|
||||
let alias = if let Some(alias) = repo.related(proxy.clone()).await? {
|
||||
alias
|
||||
} else if !config.server.read_only {
|
||||
let stream = download_stream(client, proxy.as_str(), &config).await?;
|
||||
|
||||
let (alias, _, _) = ingest_inline(stream, &repo, &store, &config).await?;
|
||||
|
||||
repo.relate_url(proxy, alias.clone()).await?;
|
||||
|
||||
alias
|
||||
} else {
|
||||
todo!("proxy URL")
|
||||
return Err(UploadError::ReadOnly.into());
|
||||
};
|
||||
|
||||
if !config.server.read_only {
|
||||
AliasAccessRepo::accessed(&repo, alias.clone()).await?;
|
||||
}
|
||||
|
||||
alias
|
||||
}
|
||||
};
|
||||
|
||||
|
@ -737,7 +773,7 @@ async fn process<R: FullRepo, S: Store + 'static>(
|
|||
};
|
||||
|
||||
if !config.server.read_only {
|
||||
repo.accessed(hash.clone(), path_string.clone()).await?;
|
||||
VariantAccessRepo::accessed(&repo, hash.clone(), path_string.clone()).await?;
|
||||
}
|
||||
|
||||
let identifier_opt = repo
|
||||
|
@ -866,7 +902,7 @@ async fn process_head<R: FullRepo, S: Store + 'static>(
|
|||
};
|
||||
|
||||
if !config.server.read_only {
|
||||
repo.accessed(hash.clone(), path_string.clone()).await?;
|
||||
VariantAccessRepo::accessed(&repo, hash.clone(), path_string.clone()).await?;
|
||||
}
|
||||
|
||||
let identifier_opt = repo
|
||||
|
@ -1002,22 +1038,37 @@ async fn do_details<R: FullRepo, S: Store + 'static>(
|
|||
}
|
||||
|
||||
/// Serve files based on alias query
|
||||
#[tracing::instrument(name = "Serving file query", skip(repo, store, config))]
|
||||
#[tracing::instrument(name = "Serving file query", skip(repo, store, client, config))]
|
||||
async fn serve_query<R: FullRepo, S: Store + 'static>(
|
||||
range: Option<web::Header<Range>>,
|
||||
web::Query(alias_query): web::Query<AliasQuery>,
|
||||
repo: web::Data<R>,
|
||||
store: web::Data<S>,
|
||||
client: web::Data<ClientWithMiddleware>,
|
||||
config: web::Data<Configuration>,
|
||||
) -> Result<HttpResponse, Error> {
|
||||
let alias = match alias_query {
|
||||
AliasQuery::Alias { alias } => Serde::into_inner(alias),
|
||||
AliasQuery::Proxy { proxy } => {
|
||||
if let Some(alias) = repo.related(proxy).await? {
|
||||
let alias = if let Some(alias) = repo.related(proxy.clone()).await? {
|
||||
alias
|
||||
} else if !config.server.read_only {
|
||||
let stream = download_stream(client, proxy.as_str(), &config).await?;
|
||||
|
||||
let (alias, _, _) = ingest_inline(stream, &repo, &store, &config).await?;
|
||||
|
||||
repo.relate_url(proxy, alias.clone()).await?;
|
||||
|
||||
alias
|
||||
} else {
|
||||
todo!("proxy URL")
|
||||
return Err(UploadError::ReadOnly.into());
|
||||
};
|
||||
|
||||
if !config.server.read_only {
|
||||
AliasAccessRepo::accessed(&repo, alias.clone()).await?;
|
||||
}
|
||||
|
||||
alias
|
||||
}
|
||||
};
|
||||
|
||||
|
@ -1288,6 +1339,7 @@ async fn clean_variants<R: FullRepo>(
|
|||
}
|
||||
|
||||
#[derive(Debug, serde::Deserialize)]
|
||||
#[serde(untagged)]
|
||||
enum AliasQuery {
|
||||
Proxy { proxy: url::Url },
|
||||
Alias { alias: Serde<Alias> },
|
||||
|
@ -1297,6 +1349,7 @@ enum AliasQuery {
|
|||
async fn set_not_found<R: FullRepo>(
|
||||
json: web::Json<AliasQuery>,
|
||||
repo: web::Data<R>,
|
||||
client: web::Data<ClientWithMiddleware>,
|
||||
config: web::Data<Configuration>,
|
||||
) -> Result<HttpResponse, Error> {
|
||||
if config.server.read_only {
|
||||
|
@ -1305,12 +1358,10 @@ async fn set_not_found<R: FullRepo>(
|
|||
|
||||
let alias = match json.into_inner() {
|
||||
AliasQuery::Alias { alias } => Serde::into_inner(alias),
|
||||
AliasQuery::Proxy { proxy } => {
|
||||
if let Some(alias) = repo.related(proxy).await? {
|
||||
alias
|
||||
} else {
|
||||
todo!("proxy URL")
|
||||
}
|
||||
AliasQuery::Proxy { .. } => {
|
||||
return Ok(HttpResponse::BadRequest().json(serde_json::json!({
|
||||
"msg": "Cannot use proxied media as Not Found image",
|
||||
})));
|
||||
}
|
||||
};
|
||||
|
||||
|
@ -1567,6 +1618,13 @@ where
|
|||
format!("\n{e}\n{e:?}")
|
||||
);
|
||||
}
|
||||
|
||||
if let Err(e) = queue::cleanup_outdated_proxies(&repo).await {
|
||||
tracing::warn!(
|
||||
"Failed to spawn cleanup for outdated proxies:{}",
|
||||
format!("\n{e}\n{e:?}")
|
||||
);
|
||||
}
|
||||
}
|
||||
});
|
||||
})
|
||||
|
|
|
@ -64,6 +64,7 @@ enum Cleanup {
|
|||
},
|
||||
AllVariants,
|
||||
OutdatedVariants,
|
||||
OutdatedProxies,
|
||||
}
|
||||
|
||||
#[derive(Debug, serde::Deserialize, serde::Serialize)]
|
||||
|
@ -126,6 +127,12 @@ async fn cleanup_variants<R: QueueRepo>(
|
|||
Ok(())
|
||||
}
|
||||
|
||||
pub(crate) async fn cleanup_outdated_proxies<R: QueueRepo>(repo: &R) -> Result<(), Error> {
|
||||
let job = serde_json::to_vec(&Cleanup::OutdatedProxies)?;
|
||||
repo.push(CLEANUP_QUEUE, job.into()).await?;
|
||||
Ok(())
|
||||
}
|
||||
|
||||
pub(crate) async fn cleanup_outdated_variants<R: QueueRepo>(repo: &R) -> Result<(), Error> {
|
||||
let job = serde_json::to_vec(&Cleanup::OutdatedVariants)?;
|
||||
repo.push(CLEANUP_QUEUE, job.into()).await?;
|
||||
|
|
|
@ -2,7 +2,10 @@ use crate::{
|
|||
config::Configuration,
|
||||
error::{Error, UploadError},
|
||||
queue::{Base64Bytes, Cleanup, LocalBoxFuture},
|
||||
repo::{Alias, AliasRepo, DeleteToken, FullRepo, HashRepo, IdentifierRepo, VariantAccessRepo},
|
||||
repo::{
|
||||
Alias, AliasAccessRepo, AliasRepo, DeleteToken, FullRepo, HashRepo, IdentifierRepo,
|
||||
VariantAccessRepo,
|
||||
},
|
||||
serde_str::Serde,
|
||||
store::{Identifier, Store},
|
||||
};
|
||||
|
@ -44,6 +47,7 @@ where
|
|||
} => hash_variant::<R, S>(repo, hash, variant).await?,
|
||||
Cleanup::AllVariants => all_variants::<R, S>(repo).await?,
|
||||
Cleanup::OutdatedVariants => outdated_variants::<R, S>(repo, configuration).await?,
|
||||
Cleanup::OutdatedProxies => outdated_proxies::<R, S>(repo, configuration).await?,
|
||||
},
|
||||
Err(e) => {
|
||||
tracing::warn!("Invalid job: {}", format!("{e}"));
|
||||
|
@ -139,6 +143,8 @@ where
|
|||
let hash = repo.hash(&alias).await?;
|
||||
|
||||
AliasRepo::cleanup(repo, &alias).await?;
|
||||
repo.remove_relation(alias.clone()).await?;
|
||||
AliasAccessRepo::remove_access(repo, alias.clone()).await?;
|
||||
|
||||
let Some(hash) = hash else {
|
||||
// hash doesn't exist, nothing to do
|
||||
|
@ -189,6 +195,31 @@ where
|
|||
Ok(())
|
||||
}
|
||||
|
||||
#[tracing::instrument(skip_all)]
|
||||
async fn outdated_proxies<R, S>(repo: &R, config: &Configuration) -> Result<(), Error>
|
||||
where
|
||||
R: FullRepo,
|
||||
S: Store,
|
||||
{
|
||||
let now = time::OffsetDateTime::now_utc();
|
||||
let since = now.saturating_sub(config.media.retention.proxy.to_duration());
|
||||
|
||||
let mut alias_stream = Box::pin(repo.older_aliases(since).await?);
|
||||
|
||||
while let Some(res) = alias_stream.next().await {
|
||||
let alias = res?;
|
||||
if let Some(token) = repo.delete_token(&alias).await? {
|
||||
super::cleanup_alias(repo, alias, token).await?;
|
||||
} else {
|
||||
tracing::warn!("Skipping alias cleanup - no delete token");
|
||||
repo.remove_relation(alias.clone()).await?;
|
||||
AliasAccessRepo::remove_access(repo, alias).await?;
|
||||
}
|
||||
}
|
||||
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[tracing::instrument(skip_all)]
|
||||
async fn hash_variant<R, S>(
|
||||
repo: &R,
|
||||
|
|
Loading…
Reference in a new issue