mirror of
https://git.asonix.dog/asonix/pict-rs
synced 2024-11-01 18:19:56 +00:00
asonix
66e1711723
Doing this as the outermost middleware ensures all endpoints are covered. Update request deadline to turn negative deadlines into immediate failures
2215 lines
70 KiB
Rust
2215 lines
70 KiB
Rust
mod backgrounded;
|
|
mod bytes_stream;
|
|
mod concurrent_processor;
|
|
mod config;
|
|
mod details;
|
|
mod discover;
|
|
mod either;
|
|
mod error;
|
|
mod error_code;
|
|
mod exiftool;
|
|
mod ffmpeg;
|
|
mod file;
|
|
mod formats;
|
|
mod future;
|
|
mod generate;
|
|
mod ingest;
|
|
mod init_tracing;
|
|
mod magick;
|
|
mod middleware;
|
|
mod migrate_store;
|
|
mod process;
|
|
mod processor;
|
|
mod queue;
|
|
mod range;
|
|
mod read;
|
|
mod repo;
|
|
mod repo_04;
|
|
mod serde_str;
|
|
mod store;
|
|
mod stream;
|
|
mod sync;
|
|
mod tmp_file;
|
|
mod validate;
|
|
|
|
use actix_form_data::{Field, Form, FormData, Multipart, Value};
|
|
use actix_web::{
|
|
guard,
|
|
http::header::{CacheControl, CacheDirective, LastModified, Range, ACCEPT_RANGES},
|
|
web, App, HttpRequest, HttpResponse, HttpResponseBuilder, HttpServer,
|
|
};
|
|
use details::{ApiDetails, HumanDate};
|
|
use future::WithTimeout;
|
|
use futures_core::Stream;
|
|
use metrics_exporter_prometheus::PrometheusBuilder;
|
|
use middleware::{Metrics, Payload};
|
|
use repo::ArcRepo;
|
|
use reqwest_middleware::{ClientBuilder, ClientWithMiddleware};
|
|
use reqwest_tracing::TracingMiddleware;
|
|
use rusty_s3::UrlStyle;
|
|
use std::{
|
|
marker::PhantomData,
|
|
path::Path,
|
|
path::PathBuf,
|
|
sync::{Arc, OnceLock},
|
|
time::{Duration, SystemTime},
|
|
};
|
|
use streem::IntoStreamer;
|
|
use tokio::sync::Semaphore;
|
|
use tracing::Instrument;
|
|
use tracing_actix_web::TracingLogger;
|
|
|
|
use self::{
|
|
backgrounded::Backgrounded,
|
|
concurrent_processor::ProcessMap,
|
|
config::{Configuration, Operation},
|
|
details::Details,
|
|
either::Either,
|
|
error::{Error, UploadError},
|
|
formats::InputProcessableFormat,
|
|
ingest::Session,
|
|
init_tracing::init_tracing,
|
|
middleware::{Deadline, Internal},
|
|
migrate_store::migrate_store,
|
|
queue::queue_generate,
|
|
repo::{sled::SledRepo, Alias, DeleteToken, Hash, Repo, UploadId, UploadResult},
|
|
serde_str::Serde,
|
|
store::{file_store::FileStore, object_store::ObjectStore, Store},
|
|
stream::{empty, once},
|
|
};
|
|
|
|
pub use self::config::{ConfigSource, PictRsConfiguration};
|
|
|
|
const MEGABYTES: usize = 1024 * 1024;
|
|
const MINUTES: u32 = 60;
|
|
const HOURS: u32 = 60 * MINUTES;
|
|
const DAYS: u32 = 24 * HOURS;
|
|
|
|
const NOT_FOUND_KEY: &str = "404-alias";
|
|
|
|
static PROCESS_SEMAPHORE: OnceLock<Semaphore> = OnceLock::new();
|
|
|
|
fn process_semaphore() -> &'static Semaphore {
|
|
PROCESS_SEMAPHORE.get_or_init(|| {
|
|
let permits = std::thread::available_parallelism()
|
|
.map(usize::from)
|
|
.unwrap_or(1)
|
|
.saturating_sub(1)
|
|
.max(1);
|
|
|
|
crate::sync::bare_semaphore(permits)
|
|
})
|
|
}
|
|
|
|
async fn ensure_details<S: Store + 'static>(
|
|
repo: &ArcRepo,
|
|
store: &S,
|
|
config: &Configuration,
|
|
alias: &Alias,
|
|
) -> Result<Details, Error> {
|
|
let Some(identifier) = repo.identifier_from_alias(alias).await? else {
|
|
return Err(UploadError::MissingAlias.into());
|
|
};
|
|
|
|
let details = repo.details(&identifier).await?;
|
|
|
|
if let Some(details) = details {
|
|
tracing::debug!("details exist");
|
|
Ok(details)
|
|
} else {
|
|
if config.server.read_only {
|
|
return Err(UploadError::ReadOnly.into());
|
|
}
|
|
|
|
tracing::debug!("generating new details from {:?}", identifier);
|
|
let new_details =
|
|
Details::from_store(store, &identifier, config.media.process_timeout).await?;
|
|
tracing::debug!("storing details for {:?}", identifier);
|
|
repo.relate_details(&identifier, &new_details).await?;
|
|
tracing::debug!("stored");
|
|
Ok(new_details)
|
|
}
|
|
}
|
|
|
|
struct Upload<S>(Value<Session>, PhantomData<S>);
|
|
|
|
impl<S: Store + 'static> FormData for Upload<S> {
|
|
type Item = Session;
|
|
type Error = Error;
|
|
|
|
fn form(req: &HttpRequest) -> Form<Self::Item, Self::Error> {
|
|
// Create a new Multipart Form validator
|
|
//
|
|
// This form is expecting a single array field, 'images' with at most 10 files in it
|
|
let repo = req
|
|
.app_data::<web::Data<ArcRepo>>()
|
|
.expect("No repo in request")
|
|
.clone();
|
|
let store = req
|
|
.app_data::<web::Data<S>>()
|
|
.expect("No store in request")
|
|
.clone();
|
|
let client = req
|
|
.app_data::<web::Data<ClientWithMiddleware>>()
|
|
.expect("No client in request")
|
|
.clone();
|
|
let config = req
|
|
.app_data::<web::Data<Configuration>>()
|
|
.expect("No configuration in request")
|
|
.clone();
|
|
|
|
Form::new()
|
|
.max_files(config.server.max_file_count)
|
|
.max_file_size(config.media.max_file_size * MEGABYTES)
|
|
.transform_error(transform_error)
|
|
.field(
|
|
"images",
|
|
Field::array(Field::file(move |filename, _, stream| {
|
|
let repo = repo.clone();
|
|
let store = store.clone();
|
|
let client = client.clone();
|
|
let config = config.clone();
|
|
|
|
metrics::increment_counter!("pict-rs.files", "upload" => "inline");
|
|
|
|
let span = tracing::info_span!("file-upload", ?filename);
|
|
|
|
Box::pin(
|
|
async move {
|
|
if config.server.read_only {
|
|
return Err(UploadError::ReadOnly.into());
|
|
}
|
|
|
|
let stream = crate::stream::from_err(stream);
|
|
|
|
ingest::ingest(&repo, &**store, &client, stream, None, &config.media)
|
|
.await
|
|
}
|
|
.instrument(span),
|
|
)
|
|
})),
|
|
)
|
|
}
|
|
|
|
fn extract(value: Value<Self::Item>) -> Result<Self, Self::Error> {
|
|
Ok(Upload(value, PhantomData))
|
|
}
|
|
}
|
|
|
|
struct Import<S: Store + 'static>(Value<Session>, PhantomData<S>);
|
|
|
|
impl<S: Store + 'static> FormData for Import<S> {
|
|
type Item = Session;
|
|
type Error = Error;
|
|
|
|
fn form(req: &actix_web::HttpRequest) -> Form<Self::Item, Self::Error> {
|
|
let repo = req
|
|
.app_data::<web::Data<ArcRepo>>()
|
|
.expect("No repo in request")
|
|
.clone();
|
|
let store = req
|
|
.app_data::<web::Data<S>>()
|
|
.expect("No store in request")
|
|
.clone();
|
|
let client = req
|
|
.app_data::<ClientWithMiddleware>()
|
|
.expect("No client in request")
|
|
.clone();
|
|
let config = req
|
|
.app_data::<web::Data<Configuration>>()
|
|
.expect("No configuration in request")
|
|
.clone();
|
|
|
|
// Create a new Multipart Form validator for internal imports
|
|
//
|
|
// This form is expecting a single array field, 'images' with at most 10 files in it
|
|
Form::new()
|
|
.max_files(config.server.max_file_count)
|
|
.max_file_size(config.media.max_file_size * MEGABYTES)
|
|
.transform_error(transform_error)
|
|
.field(
|
|
"images",
|
|
Field::array(Field::file(move |filename, _, stream| {
|
|
let repo = repo.clone();
|
|
let store = store.clone();
|
|
let client = client.clone();
|
|
let config = config.clone();
|
|
|
|
metrics::increment_counter!("pict-rs.files", "import" => "inline");
|
|
|
|
let span = tracing::info_span!("file-import", ?filename);
|
|
|
|
Box::pin(
|
|
async move {
|
|
if config.server.read_only {
|
|
return Err(UploadError::ReadOnly.into());
|
|
}
|
|
|
|
let stream = crate::stream::from_err(stream);
|
|
|
|
ingest::ingest(
|
|
&repo,
|
|
&**store,
|
|
&client,
|
|
stream,
|
|
Some(Alias::from_existing(&filename)),
|
|
&config.media,
|
|
)
|
|
.await
|
|
}
|
|
.instrument(span),
|
|
)
|
|
})),
|
|
)
|
|
}
|
|
|
|
fn extract(value: Value<Self::Item>) -> Result<Self, Self::Error>
|
|
where
|
|
Self: Sized,
|
|
{
|
|
Ok(Import(value, PhantomData))
|
|
}
|
|
}
|
|
|
|
/// Handle responding to successful uploads
|
|
#[tracing::instrument(name = "Uploaded files", skip(value, repo, store, config))]
|
|
async fn upload<S: Store + 'static>(
|
|
Multipart(Upload(value, _)): Multipart<Upload<S>>,
|
|
repo: web::Data<ArcRepo>,
|
|
store: web::Data<S>,
|
|
config: web::Data<Configuration>,
|
|
) -> Result<HttpResponse, Error> {
|
|
handle_upload(value, repo, store, config).await
|
|
}
|
|
|
|
/// Handle responding to successful uploads
|
|
#[tracing::instrument(name = "Imported files", skip(value, repo, store, config))]
|
|
async fn import<S: Store + 'static>(
|
|
Multipart(Import(value, _)): Multipart<Import<S>>,
|
|
repo: web::Data<ArcRepo>,
|
|
store: web::Data<S>,
|
|
config: web::Data<Configuration>,
|
|
) -> Result<HttpResponse, Error> {
|
|
handle_upload(value, repo, store, config).await
|
|
}
|
|
|
|
/// Handle responding to successful uploads
|
|
#[tracing::instrument(name = "Uploaded files", skip(value, repo, store, config))]
|
|
async fn handle_upload<S: Store + 'static>(
|
|
value: Value<Session>,
|
|
repo: web::Data<ArcRepo>,
|
|
store: web::Data<S>,
|
|
config: web::Data<Configuration>,
|
|
) -> Result<HttpResponse, Error> {
|
|
let images = value
|
|
.map()
|
|
.and_then(|mut m| m.remove("images"))
|
|
.and_then(|images| images.array())
|
|
.ok_or(UploadError::NoFiles)?;
|
|
|
|
let mut files = Vec::new();
|
|
let images = images
|
|
.into_iter()
|
|
.filter_map(|i| i.file())
|
|
.collect::<Vec<_>>();
|
|
|
|
for image in &images {
|
|
if let Some(alias) = image.result.alias() {
|
|
tracing::debug!("Uploaded {} as {:?}", image.filename, alias);
|
|
let delete_token = image.result.delete_token();
|
|
|
|
let details = ensure_details(&repo, &store, &config, alias).await?;
|
|
|
|
files.push(serde_json::json!({
|
|
"file": alias.to_string(),
|
|
"delete_token": delete_token.to_string(),
|
|
"details": details.into_api_details(),
|
|
}));
|
|
}
|
|
}
|
|
|
|
for image in images {
|
|
image.result.disarm();
|
|
}
|
|
|
|
Ok(HttpResponse::Created().json(&serde_json::json!({
|
|
"msg": "ok",
|
|
"files": files
|
|
})))
|
|
}
|
|
|
|
struct BackgroundedUpload<S: Store + 'static>(Value<Backgrounded>, PhantomData<S>);
|
|
|
|
impl<S: Store + 'static> FormData for BackgroundedUpload<S> {
|
|
type Item = Backgrounded;
|
|
type Error = Error;
|
|
|
|
fn form(req: &actix_web::HttpRequest) -> Form<Self::Item, Self::Error> {
|
|
// Create a new Multipart Form validator for backgrounded uploads
|
|
//
|
|
// This form is expecting a single array field, 'images' with at most 10 files in it
|
|
let repo = req
|
|
.app_data::<web::Data<ArcRepo>>()
|
|
.expect("No repo in request")
|
|
.clone();
|
|
let store = req
|
|
.app_data::<web::Data<S>>()
|
|
.expect("No store in request")
|
|
.clone();
|
|
let config = req
|
|
.app_data::<web::Data<Configuration>>()
|
|
.expect("No configuration in request")
|
|
.clone();
|
|
|
|
let read_only = config.server.read_only;
|
|
|
|
Form::new()
|
|
.max_files(config.server.max_file_count)
|
|
.max_file_size(config.media.max_file_size * MEGABYTES)
|
|
.transform_error(transform_error)
|
|
.field(
|
|
"images",
|
|
Field::array(Field::file(move |filename, _, stream| {
|
|
let repo = (**repo).clone();
|
|
let store = (**store).clone();
|
|
|
|
metrics::increment_counter!("pict-rs.files", "upload" => "background");
|
|
|
|
let span = tracing::info_span!("file-proxy", ?filename);
|
|
|
|
Box::pin(
|
|
async move {
|
|
if read_only {
|
|
return Err(UploadError::ReadOnly.into());
|
|
}
|
|
|
|
let stream = crate::stream::from_err(stream);
|
|
|
|
Backgrounded::proxy(repo, store, stream).await
|
|
}
|
|
.instrument(span),
|
|
)
|
|
})),
|
|
)
|
|
}
|
|
|
|
fn extract(value: Value<Self::Item>) -> Result<Self, Self::Error>
|
|
where
|
|
Self: Sized,
|
|
{
|
|
Ok(BackgroundedUpload(value, PhantomData))
|
|
}
|
|
}
|
|
|
|
#[tracing::instrument(name = "Uploaded files", skip(value, repo))]
|
|
async fn upload_backgrounded<S: Store>(
|
|
Multipart(BackgroundedUpload(value, _)): Multipart<BackgroundedUpload<S>>,
|
|
repo: web::Data<ArcRepo>,
|
|
) -> Result<HttpResponse, Error> {
|
|
let images = value
|
|
.map()
|
|
.and_then(|mut m| m.remove("images"))
|
|
.and_then(|images| images.array())
|
|
.ok_or(UploadError::NoFiles)?;
|
|
|
|
let mut files = Vec::new();
|
|
let images = images
|
|
.into_iter()
|
|
.filter_map(|i| i.file())
|
|
.collect::<Vec<_>>();
|
|
|
|
for image in &images {
|
|
let upload_id = image.result.upload_id().expect("Upload ID exists");
|
|
let identifier = image.result.identifier().expect("Identifier exists");
|
|
|
|
queue::queue_ingest(&repo, identifier, upload_id, None).await?;
|
|
|
|
files.push(serde_json::json!({
|
|
"upload_id": upload_id.to_string(),
|
|
}));
|
|
}
|
|
|
|
for image in images {
|
|
image.result.disarm();
|
|
}
|
|
|
|
Ok(HttpResponse::Accepted().json(&serde_json::json!({
|
|
"msg": "ok",
|
|
"uploads": files
|
|
})))
|
|
}
|
|
|
|
#[derive(Debug, serde::Deserialize)]
|
|
struct ClaimQuery {
|
|
upload_id: Serde<UploadId>,
|
|
}
|
|
|
|
/// Claim a backgrounded upload
|
|
#[tracing::instrument(name = "Waiting on upload", skip_all)]
|
|
async fn claim_upload<S: Store + 'static>(
|
|
repo: web::Data<ArcRepo>,
|
|
store: web::Data<S>,
|
|
config: web::Data<Configuration>,
|
|
query: web::Query<ClaimQuery>,
|
|
) -> Result<HttpResponse, Error> {
|
|
let upload_id = Serde::into_inner(query.into_inner().upload_id);
|
|
|
|
match repo
|
|
.wait(upload_id)
|
|
.with_timeout(Duration::from_secs(10))
|
|
.await
|
|
{
|
|
Ok(wait_res) => {
|
|
let upload_result = wait_res?;
|
|
repo.claim(upload_id).await?;
|
|
metrics::increment_counter!("pict-rs.background.upload.claim");
|
|
|
|
match upload_result {
|
|
UploadResult::Success { alias, token } => {
|
|
let details = ensure_details(&repo, &store, &config, &alias).await?;
|
|
|
|
Ok(HttpResponse::Ok().json(&serde_json::json!({
|
|
"msg": "ok",
|
|
"files": [{
|
|
"file": alias.to_string(),
|
|
"delete_token": token.to_string(),
|
|
"details": details.into_api_details(),
|
|
}]
|
|
})))
|
|
}
|
|
UploadResult::Failure { message, code } => Ok(HttpResponse::UnprocessableEntity()
|
|
.json(&serde_json::json!({
|
|
"msg": message,
|
|
"code": code,
|
|
}))),
|
|
}
|
|
}
|
|
Err(_) => Ok(HttpResponse::NoContent().finish()),
|
|
}
|
|
}
|
|
|
|
#[derive(Debug, serde::Deserialize)]
|
|
struct UrlQuery {
|
|
url: String,
|
|
|
|
#[serde(default)]
|
|
backgrounded: bool,
|
|
}
|
|
|
|
async fn ingest_inline<S: Store + 'static>(
|
|
stream: impl Stream<Item = Result<web::Bytes, Error>> + 'static,
|
|
repo: &ArcRepo,
|
|
store: &S,
|
|
client: &ClientWithMiddleware,
|
|
config: &Configuration,
|
|
) -> Result<(Alias, DeleteToken, Details), Error> {
|
|
let session = ingest::ingest(repo, store, client, stream, None, &config.media).await?;
|
|
|
|
let alias = session.alias().expect("alias should exist").to_owned();
|
|
|
|
let details = ensure_details(repo, store, config, &alias).await?;
|
|
|
|
let delete_token = session.disarm();
|
|
|
|
Ok((alias, delete_token, details))
|
|
}
|
|
|
|
/// download an image from a URL
|
|
#[tracing::instrument(name = "Downloading file", skip(client, repo, store, config))]
|
|
async fn download<S: Store + 'static>(
|
|
client: web::Data<ClientWithMiddleware>,
|
|
repo: web::Data<ArcRepo>,
|
|
store: web::Data<S>,
|
|
config: web::Data<Configuration>,
|
|
query: web::Query<UrlQuery>,
|
|
) -> Result<HttpResponse, Error> {
|
|
let stream = download_stream(&client, &query.url, &config).await?;
|
|
|
|
if query.backgrounded {
|
|
do_download_backgrounded(stream, repo, store).await
|
|
} else {
|
|
do_download_inline(stream, repo, store, &client, config).await
|
|
}
|
|
}
|
|
|
|
async fn download_stream(
|
|
client: &ClientWithMiddleware,
|
|
url: &str,
|
|
config: &Configuration,
|
|
) -> Result<impl Stream<Item = Result<web::Bytes, Error>> + 'static, Error> {
|
|
if config.server.read_only {
|
|
return Err(UploadError::ReadOnly.into());
|
|
}
|
|
|
|
let res = client.get(url).send().await?;
|
|
|
|
if !res.status().is_success() {
|
|
return Err(UploadError::Download(res.status()).into());
|
|
}
|
|
|
|
let stream = crate::stream::limit(
|
|
config.media.max_file_size * MEGABYTES,
|
|
crate::stream::from_err(res.bytes_stream()),
|
|
);
|
|
|
|
Ok(stream)
|
|
}
|
|
|
|
#[tracing::instrument(
|
|
name = "Downloading file inline",
|
|
skip(stream, repo, store, client, config)
|
|
)]
|
|
async fn do_download_inline<S: Store + 'static>(
|
|
stream: impl Stream<Item = Result<web::Bytes, Error>> + 'static,
|
|
repo: web::Data<ArcRepo>,
|
|
store: web::Data<S>,
|
|
client: &ClientWithMiddleware,
|
|
config: web::Data<Configuration>,
|
|
) -> Result<HttpResponse, Error> {
|
|
metrics::increment_counter!("pict-rs.files", "download" => "inline");
|
|
|
|
let (alias, delete_token, details) =
|
|
ingest_inline(stream, &repo, &store, client, &config).await?;
|
|
|
|
Ok(HttpResponse::Created().json(&serde_json::json!({
|
|
"msg": "ok",
|
|
"files": [{
|
|
"file": alias.to_string(),
|
|
"delete_token": delete_token.to_string(),
|
|
"details": details.into_api_details(),
|
|
}]
|
|
})))
|
|
}
|
|
|
|
#[tracing::instrument(name = "Downloading file in background", skip(stream, repo, store))]
|
|
async fn do_download_backgrounded<S: Store + 'static>(
|
|
stream: impl Stream<Item = Result<web::Bytes, Error>> + 'static,
|
|
repo: web::Data<ArcRepo>,
|
|
store: web::Data<S>,
|
|
) -> Result<HttpResponse, Error> {
|
|
metrics::increment_counter!("pict-rs.files", "download" => "background");
|
|
|
|
let backgrounded = Backgrounded::proxy((**repo).clone(), (**store).clone(), stream).await?;
|
|
|
|
let upload_id = backgrounded.upload_id().expect("Upload ID exists");
|
|
let identifier = backgrounded.identifier().expect("Identifier exists");
|
|
|
|
queue::queue_ingest(&repo, identifier, upload_id, None).await?;
|
|
|
|
backgrounded.disarm();
|
|
|
|
Ok(HttpResponse::Accepted().json(&serde_json::json!({
|
|
"msg": "ok",
|
|
"uploads": [{
|
|
"upload_id": upload_id.to_string(),
|
|
}]
|
|
})))
|
|
}
|
|
|
|
#[derive(Debug, serde::Deserialize)]
|
|
struct PageQuery {
|
|
slug: Option<String>,
|
|
timestamp: Option<HumanDate>,
|
|
limit: Option<usize>,
|
|
}
|
|
|
|
#[derive(serde::Serialize)]
|
|
struct PageJson {
|
|
limit: usize,
|
|
|
|
#[serde(skip_serializing_if = "Option::is_none")]
|
|
current: Option<String>,
|
|
|
|
#[serde(skip_serializing_if = "Option::is_none")]
|
|
prev: Option<String>,
|
|
|
|
#[serde(skip_serializing_if = "Option::is_none")]
|
|
next: Option<String>,
|
|
|
|
hashes: Vec<HashJson>,
|
|
}
|
|
|
|
#[derive(serde::Serialize)]
|
|
struct HashJson {
|
|
hex: String,
|
|
aliases: Vec<String>,
|
|
details: Option<ApiDetails>,
|
|
}
|
|
|
|
/// Get a page of hashes
|
|
#[tracing::instrument(name = "Hash Page", skip(repo))]
|
|
async fn page(
|
|
repo: web::Data<ArcRepo>,
|
|
web::Query(PageQuery {
|
|
slug,
|
|
timestamp,
|
|
limit,
|
|
}): web::Query<PageQuery>,
|
|
) -> Result<HttpResponse, Error> {
|
|
let limit = limit.unwrap_or(20);
|
|
|
|
let page = if let Some(timestamp) = timestamp {
|
|
repo.hash_page_by_date(timestamp.timestamp, limit).await?
|
|
} else {
|
|
repo.hash_page(slug, limit).await?
|
|
};
|
|
|
|
let mut hashes = Vec::with_capacity(page.hashes.len());
|
|
|
|
for hash in &page.hashes {
|
|
let hex = hash.to_hex();
|
|
let aliases = repo
|
|
.aliases_for_hash(hash.clone())
|
|
.await?
|
|
.into_iter()
|
|
.map(|a| a.to_string())
|
|
.collect();
|
|
|
|
let identifier = repo.identifier(hash.clone()).await?;
|
|
let details = if let Some(identifier) = identifier {
|
|
repo.details(&identifier)
|
|
.await?
|
|
.map(|d| d.into_api_details())
|
|
} else {
|
|
None
|
|
};
|
|
|
|
hashes.push(HashJson {
|
|
hex,
|
|
aliases,
|
|
details,
|
|
});
|
|
}
|
|
|
|
let page = PageJson {
|
|
limit: page.limit,
|
|
current: page.current(),
|
|
prev: page.prev(),
|
|
next: page.next(),
|
|
hashes,
|
|
};
|
|
|
|
Ok(HttpResponse::Ok().json(serde_json::json!({
|
|
"msg": "ok",
|
|
"page": page,
|
|
})))
|
|
}
|
|
|
|
/// Delete aliases and files
|
|
#[tracing::instrument(name = "Deleting file", skip(repo, config))]
|
|
async fn delete(
|
|
repo: web::Data<ArcRepo>,
|
|
config: web::Data<Configuration>,
|
|
path_entries: web::Path<(String, String)>,
|
|
) -> Result<HttpResponse, Error> {
|
|
if config.server.read_only {
|
|
return Err(UploadError::ReadOnly.into());
|
|
}
|
|
|
|
let (token, alias) = path_entries.into_inner();
|
|
|
|
let token = DeleteToken::from_existing(&token);
|
|
let alias = Alias::from_existing(&alias);
|
|
|
|
queue::cleanup_alias(&repo, alias, token).await?;
|
|
|
|
Ok(HttpResponse::NoContent().finish())
|
|
}
|
|
|
|
#[derive(Debug, serde::Deserialize, serde::Serialize, PartialEq, Eq, PartialOrd, Ord)]
|
|
#[serde(untagged)]
|
|
enum ProcessSource {
|
|
Source { src: Serde<Alias> },
|
|
Alias { alias: Serde<Alias> },
|
|
Proxy { proxy: url::Url },
|
|
}
|
|
|
|
#[derive(Debug, serde::Deserialize, serde::Serialize, PartialEq, Eq, PartialOrd, Ord)]
|
|
struct ProcessQuery {
|
|
#[serde(flatten)]
|
|
source: ProcessSource,
|
|
|
|
#[serde(with = "tuple_vec_map", flatten)]
|
|
operations: Vec<(String, String)>,
|
|
}
|
|
|
|
fn prepare_process(
|
|
config: &Configuration,
|
|
operations: Vec<(String, String)>,
|
|
ext: &str,
|
|
) -> Result<(InputProcessableFormat, PathBuf, Vec<String>), Error> {
|
|
let operations = operations
|
|
.into_iter()
|
|
.filter(|(k, _)| config.media.filters.contains(&k.to_lowercase()))
|
|
.collect::<Vec<_>>();
|
|
|
|
let format = ext
|
|
.parse::<InputProcessableFormat>()
|
|
.map_err(|_| UploadError::UnsupportedProcessExtension)?;
|
|
|
|
let (thumbnail_path, thumbnail_args) =
|
|
self::processor::build_chain(&operations, &format.to_string())?;
|
|
|
|
Ok((format, thumbnail_path, thumbnail_args))
|
|
}
|
|
|
|
#[tracing::instrument(name = "Fetching derived details", skip(repo, config))]
|
|
async fn process_details<S: Store>(
|
|
web::Query(ProcessQuery { source, operations }): web::Query<ProcessQuery>,
|
|
ext: web::Path<String>,
|
|
repo: web::Data<ArcRepo>,
|
|
config: web::Data<Configuration>,
|
|
) -> Result<HttpResponse, Error> {
|
|
let alias = match source {
|
|
ProcessSource::Alias { alias } | ProcessSource::Source { src: alias } => {
|
|
Serde::into_inner(alias)
|
|
}
|
|
ProcessSource::Proxy { proxy } => {
|
|
let Some(alias) = repo.related(proxy).await? else {
|
|
return Ok(HttpResponse::NotFound().json(&serde_json::json!({
|
|
"msg": "No images associated with provided proxy url"
|
|
})));
|
|
};
|
|
alias
|
|
}
|
|
};
|
|
|
|
let (_, thumbnail_path, _) = prepare_process(&config, operations, ext.as_str())?;
|
|
|
|
let Some(hash) = repo.hash(&alias).await? else {
|
|
// Invalid alias
|
|
return Ok(HttpResponse::NotFound().json(&serde_json::json!({
|
|
"msg": "No images associated with provided alias",
|
|
})));
|
|
};
|
|
|
|
let thumbnail_string = thumbnail_path.to_string_lossy().to_string();
|
|
|
|
if !config.server.read_only {
|
|
repo.accessed_variant(hash.clone(), thumbnail_string.clone())
|
|
.await?;
|
|
}
|
|
|
|
let identifier = repo
|
|
.variant_identifier(hash, thumbnail_string)
|
|
.await?
|
|
.ok_or(UploadError::MissingAlias)?;
|
|
|
|
let details = repo.details(&identifier).await?;
|
|
|
|
let details = details.ok_or(UploadError::NoFiles)?;
|
|
|
|
Ok(HttpResponse::Ok().json(&details.into_api_details()))
|
|
}
|
|
|
|
async fn not_found_hash(repo: &ArcRepo) -> Result<Option<(Alias, Hash)>, Error> {
|
|
let Some(not_found) = repo.get(NOT_FOUND_KEY).await? else {
|
|
return Ok(None);
|
|
};
|
|
|
|
let Some(alias) = Alias::from_slice(not_found.as_ref()) else {
|
|
tracing::warn!("Couldn't parse not-found alias");
|
|
return Ok(None);
|
|
};
|
|
|
|
let Some(hash) = repo.hash(&alias).await? else {
|
|
tracing::warn!("No hash found for not-found alias");
|
|
return Ok(None);
|
|
};
|
|
|
|
Ok(Some((alias, hash)))
|
|
}
|
|
|
|
/// Process files
|
|
#[allow(clippy::too_many_arguments)]
|
|
#[tracing::instrument(
|
|
name = "Serving processed image",
|
|
skip(repo, store, client, config, process_map)
|
|
)]
|
|
async fn process<S: Store + 'static>(
|
|
range: Option<web::Header<Range>>,
|
|
web::Query(ProcessQuery { source, operations }): web::Query<ProcessQuery>,
|
|
ext: web::Path<String>,
|
|
repo: web::Data<ArcRepo>,
|
|
store: web::Data<S>,
|
|
client: web::Data<ClientWithMiddleware>,
|
|
config: web::Data<Configuration>,
|
|
process_map: web::Data<ProcessMap>,
|
|
) -> Result<HttpResponse, Error> {
|
|
let alias = match source {
|
|
ProcessSource::Alias { alias } | ProcessSource::Source { src: alias } => {
|
|
Serde::into_inner(alias)
|
|
}
|
|
ProcessSource::Proxy { proxy } => {
|
|
let alias = if let Some(alias) = repo.related(proxy.clone()).await? {
|
|
alias
|
|
} else if !config.server.read_only {
|
|
let stream = download_stream(&client, proxy.as_str(), &config).await?;
|
|
|
|
let (alias, _, _) = ingest_inline(stream, &repo, &store, &client, &config).await?;
|
|
|
|
repo.relate_url(proxy, alias.clone()).await?;
|
|
|
|
alias
|
|
} else {
|
|
return Err(UploadError::ReadOnly.into());
|
|
};
|
|
|
|
if !config.server.read_only {
|
|
repo.accessed_alias(alias.clone()).await?;
|
|
}
|
|
|
|
alias
|
|
}
|
|
};
|
|
|
|
let (format, thumbnail_path, thumbnail_args) =
|
|
prepare_process(&config, operations, ext.as_str())?;
|
|
|
|
let path_string = thumbnail_path.to_string_lossy().to_string();
|
|
|
|
let (hash, alias, not_found) = if let Some(hash) = repo.hash(&alias).await? {
|
|
(hash, alias, false)
|
|
} else {
|
|
let Some((alias, hash)) = not_found_hash(&repo).await? else {
|
|
return Ok(HttpResponse::NotFound().finish());
|
|
};
|
|
|
|
(hash, alias, true)
|
|
};
|
|
|
|
if !config.server.read_only {
|
|
repo.accessed_variant(hash.clone(), path_string.clone())
|
|
.await?;
|
|
}
|
|
|
|
let identifier_opt = repo.variant_identifier(hash.clone(), path_string).await?;
|
|
|
|
if let Some(identifier) = identifier_opt {
|
|
let details = repo.details(&identifier).await?;
|
|
|
|
let details = if let Some(details) = details {
|
|
tracing::debug!("details exist");
|
|
details
|
|
} else {
|
|
if config.server.read_only {
|
|
return Err(UploadError::ReadOnly.into());
|
|
}
|
|
|
|
tracing::debug!("generating new details from {:?}", identifier);
|
|
let new_details =
|
|
Details::from_store(&store, &identifier, config.media.process_timeout).await?;
|
|
tracing::debug!("storing details for {:?}", identifier);
|
|
repo.relate_details(&identifier, &new_details).await?;
|
|
tracing::debug!("stored");
|
|
new_details
|
|
};
|
|
|
|
if let Some(public_url) = store.public_url(&identifier) {
|
|
return Ok(HttpResponse::SeeOther()
|
|
.insert_header((actix_web::http::header::LOCATION, public_url.as_str()))
|
|
.finish());
|
|
}
|
|
|
|
return ranged_file_resp(&store, identifier, range, details, not_found).await;
|
|
}
|
|
|
|
if config.server.read_only {
|
|
return Err(UploadError::ReadOnly.into());
|
|
}
|
|
|
|
let original_details = ensure_details(&repo, &store, &config, &alias).await?;
|
|
|
|
let (details, bytes) = generate::generate(
|
|
&repo,
|
|
&store,
|
|
&process_map,
|
|
format,
|
|
thumbnail_path,
|
|
thumbnail_args,
|
|
&original_details,
|
|
&config.media,
|
|
hash,
|
|
)
|
|
.await?;
|
|
|
|
let (builder, stream) = if let Some(web::Header(range_header)) = range {
|
|
if let Some(range) = range::single_bytes_range(&range_header) {
|
|
let len = bytes.len() as u64;
|
|
|
|
if let Some(content_range) = range::to_content_range(range, len) {
|
|
let mut builder = HttpResponse::PartialContent();
|
|
builder.insert_header(content_range);
|
|
let stream = range::chop_bytes(range, bytes, len)?;
|
|
|
|
(builder, Either::left(Either::left(stream)))
|
|
} else {
|
|
(
|
|
HttpResponse::RangeNotSatisfiable(),
|
|
Either::left(Either::right(empty())),
|
|
)
|
|
}
|
|
} else {
|
|
return Err(UploadError::Range.into());
|
|
}
|
|
} else if not_found {
|
|
(HttpResponse::NotFound(), Either::right(once(Ok(bytes))))
|
|
} else {
|
|
(HttpResponse::Ok(), Either::right(once(Ok(bytes))))
|
|
};
|
|
|
|
Ok(srv_response(
|
|
builder,
|
|
stream,
|
|
details.media_type(),
|
|
7 * DAYS,
|
|
details.system_time(),
|
|
))
|
|
}
|
|
|
|
#[tracing::instrument(name = "Serving processed image headers", skip(repo, store, config))]
|
|
async fn process_head<S: Store + 'static>(
|
|
range: Option<web::Header<Range>>,
|
|
web::Query(ProcessQuery { source, operations }): web::Query<ProcessQuery>,
|
|
ext: web::Path<String>,
|
|
repo: web::Data<ArcRepo>,
|
|
store: web::Data<S>,
|
|
config: web::Data<Configuration>,
|
|
) -> Result<HttpResponse, Error> {
|
|
let alias = match source {
|
|
ProcessSource::Alias { alias } | ProcessSource::Source { src: alias } => {
|
|
Serde::into_inner(alias)
|
|
}
|
|
ProcessSource::Proxy { proxy } => {
|
|
let Some(alias) = repo.related(proxy).await? else {
|
|
return Ok(HttpResponse::NotFound().finish());
|
|
};
|
|
alias
|
|
}
|
|
};
|
|
|
|
let (_, thumbnail_path, _) = prepare_process(&config, operations, ext.as_str())?;
|
|
|
|
let path_string = thumbnail_path.to_string_lossy().to_string();
|
|
let Some(hash) = repo.hash(&alias).await? else {
|
|
// Invalid alias
|
|
return Ok(HttpResponse::NotFound().finish());
|
|
};
|
|
|
|
if !config.server.read_only {
|
|
repo.accessed_variant(hash.clone(), path_string.clone())
|
|
.await?;
|
|
}
|
|
|
|
let identifier_opt = repo.variant_identifier(hash.clone(), path_string).await?;
|
|
|
|
if let Some(identifier) = identifier_opt {
|
|
let details = repo.details(&identifier).await?;
|
|
|
|
let details = if let Some(details) = details {
|
|
tracing::debug!("details exist");
|
|
details
|
|
} else {
|
|
if config.server.read_only {
|
|
return Err(UploadError::ReadOnly.into());
|
|
}
|
|
|
|
tracing::debug!("generating new details from {:?}", identifier);
|
|
let new_details =
|
|
Details::from_store(&store, &identifier, config.media.process_timeout).await?;
|
|
tracing::debug!("storing details for {:?}", identifier);
|
|
repo.relate_details(&identifier, &new_details).await?;
|
|
tracing::debug!("stored");
|
|
new_details
|
|
};
|
|
|
|
if let Some(public_url) = store.public_url(&identifier) {
|
|
return Ok(HttpResponse::SeeOther()
|
|
.insert_header((actix_web::http::header::LOCATION, public_url.as_str()))
|
|
.finish());
|
|
}
|
|
|
|
return ranged_file_head_resp(&store, identifier, range, details).await;
|
|
}
|
|
|
|
Ok(HttpResponse::NotFound().finish())
|
|
}
|
|
|
|
/// Process files
|
|
#[tracing::instrument(name = "Spawning image process", skip(repo))]
|
|
async fn process_backgrounded<S: Store>(
|
|
web::Query(ProcessQuery { source, operations }): web::Query<ProcessQuery>,
|
|
ext: web::Path<String>,
|
|
repo: web::Data<ArcRepo>,
|
|
config: web::Data<Configuration>,
|
|
) -> Result<HttpResponse, Error> {
|
|
let source = match source {
|
|
ProcessSource::Alias { alias } | ProcessSource::Source { src: alias } => {
|
|
Serde::into_inner(alias)
|
|
}
|
|
ProcessSource::Proxy { proxy } => {
|
|
let Some(alias) = repo.related(proxy).await? else {
|
|
return Ok(HttpResponse::NotFound().finish());
|
|
};
|
|
alias
|
|
}
|
|
};
|
|
|
|
let (target_format, process_path, process_args) =
|
|
prepare_process(&config, operations, ext.as_str())?;
|
|
|
|
let path_string = process_path.to_string_lossy().to_string();
|
|
let Some(hash) = repo.hash(&source).await? else {
|
|
// Invalid alias
|
|
return Ok(HttpResponse::BadRequest().finish());
|
|
};
|
|
|
|
let identifier_opt = repo.variant_identifier(hash.clone(), path_string).await?;
|
|
|
|
if identifier_opt.is_some() {
|
|
return Ok(HttpResponse::Accepted().finish());
|
|
}
|
|
|
|
if config.server.read_only {
|
|
return Err(UploadError::ReadOnly.into());
|
|
}
|
|
|
|
queue_generate(&repo, target_format, source, process_path, process_args).await?;
|
|
|
|
Ok(HttpResponse::Accepted().finish())
|
|
}
|
|
|
|
/// Fetch file details
|
|
#[tracing::instrument(name = "Fetching query details", skip(repo, store, config))]
|
|
async fn details_query<S: Store + 'static>(
|
|
web::Query(alias_query): web::Query<AliasQuery>,
|
|
repo: web::Data<ArcRepo>,
|
|
store: web::Data<S>,
|
|
config: web::Data<Configuration>,
|
|
) -> Result<HttpResponse, Error> {
|
|
let alias = match alias_query {
|
|
AliasQuery::Alias { alias } => Serde::into_inner(alias),
|
|
AliasQuery::Proxy { proxy } => {
|
|
let Some(alias) = repo.related(proxy).await? else {
|
|
return Ok(HttpResponse::NotFound().json(&serde_json::json!({
|
|
"msg": "Provided proxy URL has not been cached",
|
|
})));
|
|
};
|
|
alias
|
|
}
|
|
};
|
|
|
|
do_details(alias, repo, store, config).await
|
|
}
|
|
|
|
/// Fetch file details
|
|
#[tracing::instrument(name = "Fetching details", skip(repo, store, config))]
|
|
async fn details<S: Store + 'static>(
|
|
alias: web::Path<Serde<Alias>>,
|
|
repo: web::Data<ArcRepo>,
|
|
store: web::Data<S>,
|
|
config: web::Data<Configuration>,
|
|
) -> Result<HttpResponse, Error> {
|
|
do_details(Serde::into_inner(alias.into_inner()), repo, store, config).await
|
|
}
|
|
|
|
async fn do_details<S: Store + 'static>(
|
|
alias: Alias,
|
|
repo: web::Data<ArcRepo>,
|
|
store: web::Data<S>,
|
|
config: web::Data<Configuration>,
|
|
) -> Result<HttpResponse, Error> {
|
|
let details = ensure_details(&repo, &store, &config, &alias).await?;
|
|
|
|
Ok(HttpResponse::Ok().json(&details.into_api_details()))
|
|
}
|
|
|
|
/// Serve files based on alias query
|
|
#[tracing::instrument(name = "Serving file query", skip(repo, store, client, config))]
|
|
async fn serve_query<S: Store + 'static>(
|
|
range: Option<web::Header<Range>>,
|
|
web::Query(alias_query): web::Query<AliasQuery>,
|
|
repo: web::Data<ArcRepo>,
|
|
store: web::Data<S>,
|
|
client: web::Data<ClientWithMiddleware>,
|
|
config: web::Data<Configuration>,
|
|
) -> Result<HttpResponse, Error> {
|
|
let alias = match alias_query {
|
|
AliasQuery::Alias { alias } => Serde::into_inner(alias),
|
|
AliasQuery::Proxy { proxy } => {
|
|
let alias = if let Some(alias) = repo.related(proxy.clone()).await? {
|
|
alias
|
|
} else if !config.server.read_only {
|
|
let stream = download_stream(&client, proxy.as_str(), &config).await?;
|
|
|
|
let (alias, _, _) = ingest_inline(stream, &repo, &store, &client, &config).await?;
|
|
|
|
repo.relate_url(proxy, alias.clone()).await?;
|
|
|
|
alias
|
|
} else {
|
|
return Err(UploadError::ReadOnly.into());
|
|
};
|
|
|
|
if !config.server.read_only {
|
|
repo.accessed_alias(alias.clone()).await?;
|
|
}
|
|
|
|
alias
|
|
}
|
|
};
|
|
|
|
do_serve(range, alias, repo, store, config).await
|
|
}
|
|
|
|
/// Serve files
|
|
#[tracing::instrument(name = "Serving file", skip(repo, store, config))]
|
|
async fn serve<S: Store + 'static>(
|
|
range: Option<web::Header<Range>>,
|
|
alias: web::Path<Serde<Alias>>,
|
|
repo: web::Data<ArcRepo>,
|
|
store: web::Data<S>,
|
|
config: web::Data<Configuration>,
|
|
) -> Result<HttpResponse, Error> {
|
|
do_serve(
|
|
range,
|
|
Serde::into_inner(alias.into_inner()),
|
|
repo,
|
|
store,
|
|
config,
|
|
)
|
|
.await
|
|
}
|
|
|
|
async fn do_serve<S: Store + 'static>(
|
|
range: Option<web::Header<Range>>,
|
|
alias: Alias,
|
|
repo: web::Data<ArcRepo>,
|
|
store: web::Data<S>,
|
|
config: web::Data<Configuration>,
|
|
) -> Result<HttpResponse, Error> {
|
|
let (hash, alias, not_found) = if let Some(hash) = repo.hash(&alias).await? {
|
|
(hash, alias, false)
|
|
} else {
|
|
let Some((alias, hash)) = not_found_hash(&repo).await? else {
|
|
return Ok(HttpResponse::NotFound().finish());
|
|
};
|
|
|
|
(hash, alias, true)
|
|
};
|
|
|
|
let Some(identifier) = repo.identifier(hash.clone()).await? else {
|
|
tracing::warn!("Original File identifier for hash {hash:?} is missing, queue cleanup task",);
|
|
crate::queue::cleanup_hash(&repo, hash).await?;
|
|
return Ok(HttpResponse::NotFound().finish());
|
|
};
|
|
|
|
let details = ensure_details(&repo, &store, &config, &alias).await?;
|
|
|
|
if let Some(public_url) = store.public_url(&identifier) {
|
|
return Ok(HttpResponse::SeeOther()
|
|
.insert_header((actix_web::http::header::LOCATION, public_url.as_str()))
|
|
.finish());
|
|
}
|
|
|
|
ranged_file_resp(&store, identifier, range, details, not_found).await
|
|
}
|
|
|
|
#[tracing::instrument(name = "Serving query file headers", skip(repo, store, config))]
|
|
async fn serve_query_head<S: Store + 'static>(
|
|
range: Option<web::Header<Range>>,
|
|
web::Query(alias_query): web::Query<AliasQuery>,
|
|
repo: web::Data<ArcRepo>,
|
|
store: web::Data<S>,
|
|
config: web::Data<Configuration>,
|
|
) -> Result<HttpResponse, Error> {
|
|
let alias = match alias_query {
|
|
AliasQuery::Alias { alias } => Serde::into_inner(alias),
|
|
AliasQuery::Proxy { proxy } => {
|
|
let Some(alias) = repo.related(proxy).await? else {
|
|
return Ok(HttpResponse::NotFound().finish());
|
|
};
|
|
alias
|
|
}
|
|
};
|
|
|
|
do_serve_head(range, alias, repo, store, config).await
|
|
}
|
|
|
|
#[tracing::instrument(name = "Serving file headers", skip(repo, store, config))]
|
|
async fn serve_head<S: Store + 'static>(
|
|
range: Option<web::Header<Range>>,
|
|
alias: web::Path<Serde<Alias>>,
|
|
repo: web::Data<ArcRepo>,
|
|
store: web::Data<S>,
|
|
config: web::Data<Configuration>,
|
|
) -> Result<HttpResponse, Error> {
|
|
do_serve_head(
|
|
range,
|
|
Serde::into_inner(alias.into_inner()),
|
|
repo,
|
|
store,
|
|
config,
|
|
)
|
|
.await
|
|
}
|
|
|
|
async fn do_serve_head<S: Store + 'static>(
|
|
range: Option<web::Header<Range>>,
|
|
alias: Alias,
|
|
repo: web::Data<ArcRepo>,
|
|
store: web::Data<S>,
|
|
config: web::Data<Configuration>,
|
|
) -> Result<HttpResponse, Error> {
|
|
let Some(identifier) = repo.identifier_from_alias(&alias).await? else {
|
|
// Invalid alias
|
|
return Ok(HttpResponse::NotFound().finish());
|
|
};
|
|
|
|
let details = ensure_details(&repo, &store, &config, &alias).await?;
|
|
|
|
if let Some(public_url) = store.public_url(&identifier) {
|
|
return Ok(HttpResponse::SeeOther()
|
|
.insert_header((actix_web::http::header::LOCATION, public_url.as_str()))
|
|
.finish());
|
|
}
|
|
|
|
ranged_file_head_resp(&store, identifier, range, details).await
|
|
}
|
|
|
|
async fn ranged_file_head_resp<S: Store + 'static>(
|
|
store: &S,
|
|
identifier: Arc<str>,
|
|
range: Option<web::Header<Range>>,
|
|
details: Details,
|
|
) -> Result<HttpResponse, Error> {
|
|
let builder = if let Some(web::Header(range_header)) = range {
|
|
//Range header exists - return as ranged
|
|
if let Some(range) = range::single_bytes_range(&range_header) {
|
|
let len = store.len(&identifier).await?;
|
|
|
|
if let Some(content_range) = range::to_content_range(range, len) {
|
|
let mut builder = HttpResponse::PartialContent();
|
|
builder.insert_header(content_range);
|
|
builder
|
|
} else {
|
|
HttpResponse::RangeNotSatisfiable()
|
|
}
|
|
} else {
|
|
return Err(UploadError::Range.into());
|
|
}
|
|
} else {
|
|
// no range header
|
|
HttpResponse::Ok()
|
|
};
|
|
|
|
Ok(srv_head(
|
|
builder,
|
|
details.media_type(),
|
|
7 * DAYS,
|
|
details.system_time(),
|
|
)
|
|
.finish())
|
|
}
|
|
|
|
async fn ranged_file_resp<S: Store + 'static>(
|
|
store: &S,
|
|
identifier: Arc<str>,
|
|
range: Option<web::Header<Range>>,
|
|
details: Details,
|
|
not_found: bool,
|
|
) -> Result<HttpResponse, Error> {
|
|
let (builder, stream) = if let Some(web::Header(range_header)) = range {
|
|
//Range header exists - return as ranged
|
|
if let Some(range) = range::single_bytes_range(&range_header) {
|
|
let len = store.len(&identifier).await?;
|
|
|
|
if let Some(content_range) = range::to_content_range(range, len) {
|
|
let mut builder = HttpResponse::PartialContent();
|
|
builder.insert_header(content_range);
|
|
(
|
|
builder,
|
|
Either::left(Either::left(
|
|
range::chop_store(range, store, &identifier, len).await?,
|
|
)),
|
|
)
|
|
} else {
|
|
(
|
|
HttpResponse::RangeNotSatisfiable(),
|
|
Either::left(Either::right(empty())),
|
|
)
|
|
}
|
|
} else {
|
|
return Err(UploadError::Range.into());
|
|
}
|
|
} else {
|
|
//No Range header in the request - return the entire document
|
|
let stream = crate::stream::from_err(store.to_stream(&identifier, None, None).await?);
|
|
|
|
if not_found {
|
|
(HttpResponse::NotFound(), Either::right(stream))
|
|
} else {
|
|
(HttpResponse::Ok(), Either::right(stream))
|
|
}
|
|
};
|
|
|
|
Ok(srv_response(
|
|
builder,
|
|
stream,
|
|
details.media_type(),
|
|
7 * DAYS,
|
|
details.system_time(),
|
|
))
|
|
}
|
|
|
|
// A helper method to produce responses with proper cache headers
|
|
fn srv_response<S, E>(
|
|
builder: HttpResponseBuilder,
|
|
stream: S,
|
|
ext: mime::Mime,
|
|
expires: u32,
|
|
modified: SystemTime,
|
|
) -> HttpResponse
|
|
where
|
|
S: Stream<Item = Result<web::Bytes, E>> + 'static,
|
|
E: std::error::Error + 'static,
|
|
actix_web::Error: From<E>,
|
|
{
|
|
let stream = crate::stream::timeout(Duration::from_secs(5), stream);
|
|
|
|
let stream = streem::try_from_fn(|yielder| async move {
|
|
let stream = std::pin::pin!(stream);
|
|
let mut streamer = stream.into_streamer();
|
|
|
|
while let Some(res) = streamer.next().await {
|
|
let item = res.map_err(Error::from)??;
|
|
yielder.yield_ok(item).await;
|
|
}
|
|
|
|
Ok(()) as Result<(), actix_web::Error>
|
|
});
|
|
|
|
srv_head(builder, ext, expires, modified).streaming(stream)
|
|
}
|
|
|
|
// A helper method to produce responses with proper cache headers
|
|
fn srv_head(
|
|
mut builder: HttpResponseBuilder,
|
|
ext: mime::Mime,
|
|
expires: u32,
|
|
modified: SystemTime,
|
|
) -> HttpResponseBuilder {
|
|
builder
|
|
.insert_header(LastModified(modified.into()))
|
|
.insert_header(CacheControl(vec![
|
|
CacheDirective::Public,
|
|
CacheDirective::MaxAge(expires),
|
|
CacheDirective::Extension("immutable".to_owned(), None),
|
|
]))
|
|
.insert_header((ACCEPT_RANGES, "bytes"))
|
|
.content_type(ext.to_string());
|
|
|
|
builder
|
|
}
|
|
|
|
#[tracing::instrument(name = "Spawning variant cleanup", skip(repo, config))]
|
|
async fn clean_variants(
|
|
repo: web::Data<ArcRepo>,
|
|
config: web::Data<Configuration>,
|
|
) -> Result<HttpResponse, Error> {
|
|
if config.server.read_only {
|
|
return Err(UploadError::ReadOnly.into());
|
|
}
|
|
|
|
queue::cleanup_all_variants(&repo).await?;
|
|
Ok(HttpResponse::NoContent().finish())
|
|
}
|
|
|
|
#[derive(Debug, serde::Deserialize)]
|
|
#[serde(untagged)]
|
|
enum AliasQuery {
|
|
Proxy { proxy: url::Url },
|
|
Alias { alias: Serde<Alias> },
|
|
}
|
|
|
|
#[tracing::instrument(name = "Setting 404 Image", skip(repo, config))]
|
|
async fn set_not_found(
|
|
json: web::Json<AliasQuery>,
|
|
repo: web::Data<ArcRepo>,
|
|
client: web::Data<ClientWithMiddleware>,
|
|
config: web::Data<Configuration>,
|
|
) -> Result<HttpResponse, Error> {
|
|
if config.server.read_only {
|
|
return Err(UploadError::ReadOnly.into());
|
|
}
|
|
|
|
let alias = match json.into_inner() {
|
|
AliasQuery::Alias { alias } => Serde::into_inner(alias),
|
|
AliasQuery::Proxy { .. } => {
|
|
return Ok(HttpResponse::BadRequest().json(serde_json::json!({
|
|
"msg": "Cannot use proxied media as Not Found image",
|
|
})));
|
|
}
|
|
};
|
|
|
|
if repo.hash(&alias).await?.is_none() {
|
|
return Ok(HttpResponse::BadRequest().json(serde_json::json!({
|
|
"msg": "No hash associated with provided alias"
|
|
})));
|
|
}
|
|
|
|
repo.set(NOT_FOUND_KEY, alias.to_bytes().into()).await?;
|
|
|
|
Ok(HttpResponse::Created().json(serde_json::json!({
|
|
"msg": "ok",
|
|
})))
|
|
}
|
|
|
|
#[tracing::instrument(name = "Purging file", skip(repo, config))]
|
|
async fn purge(
|
|
web::Query(alias_query): web::Query<AliasQuery>,
|
|
repo: web::Data<ArcRepo>,
|
|
config: web::Data<Configuration>,
|
|
) -> Result<HttpResponse, Error> {
|
|
if config.server.read_only {
|
|
return Err(UploadError::ReadOnly.into());
|
|
}
|
|
|
|
let alias = match alias_query {
|
|
AliasQuery::Alias { alias } => Serde::into_inner(alias),
|
|
AliasQuery::Proxy { proxy } => {
|
|
let Some(alias) = repo.related(proxy).await? else {
|
|
return Ok(HttpResponse::NotFound().finish());
|
|
};
|
|
alias
|
|
}
|
|
};
|
|
|
|
let aliases = repo.aliases_from_alias(&alias).await?;
|
|
|
|
let Some(hash) = repo.hash(&alias).await? else {
|
|
return Ok(HttpResponse::BadRequest().json(&serde_json::json!({
|
|
"msg": "No images associated with provided alias",
|
|
})));
|
|
};
|
|
queue::cleanup_hash(&repo, hash).await?;
|
|
|
|
Ok(HttpResponse::Ok().json(&serde_json::json!({
|
|
"msg": "ok",
|
|
"aliases": aliases.iter().map(|a| a.to_string()).collect::<Vec<_>>()
|
|
})))
|
|
}
|
|
|
|
#[tracing::instrument(name = "Deleting alias", skip(repo, config))]
|
|
async fn delete_alias(
|
|
web::Query(alias_query): web::Query<AliasQuery>,
|
|
repo: web::Data<ArcRepo>,
|
|
config: web::Data<Configuration>,
|
|
) -> Result<HttpResponse, Error> {
|
|
if config.server.read_only {
|
|
return Err(UploadError::ReadOnly.into());
|
|
}
|
|
|
|
let alias = match alias_query {
|
|
AliasQuery::Alias { alias } => Serde::into_inner(alias),
|
|
AliasQuery::Proxy { proxy } => {
|
|
let Some(alias) = repo.related(proxy).await? else {
|
|
return Ok(HttpResponse::NotFound().finish());
|
|
};
|
|
alias
|
|
}
|
|
};
|
|
|
|
if let Some(token) = repo.delete_token(&alias).await? {
|
|
queue::cleanup_alias(&repo, alias, token).await?;
|
|
} else {
|
|
return Ok(HttpResponse::NotFound().finish());
|
|
}
|
|
|
|
Ok(HttpResponse::Ok().json(&serde_json::json!({
|
|
"msg": "ok",
|
|
})))
|
|
}
|
|
|
|
#[tracing::instrument(name = "Fetching aliases", skip(repo))]
|
|
async fn aliases(
|
|
web::Query(alias_query): web::Query<AliasQuery>,
|
|
repo: web::Data<ArcRepo>,
|
|
) -> Result<HttpResponse, Error> {
|
|
let alias = match alias_query {
|
|
AliasQuery::Alias { alias } => Serde::into_inner(alias),
|
|
AliasQuery::Proxy { proxy } => {
|
|
let Some(alias) = repo.related(proxy).await? else {
|
|
return Ok(HttpResponse::NotFound().finish());
|
|
};
|
|
alias
|
|
}
|
|
};
|
|
|
|
let aliases = repo.aliases_from_alias(&alias).await?;
|
|
|
|
Ok(HttpResponse::Ok().json(&serde_json::json!({
|
|
"msg": "ok",
|
|
"aliases": aliases.iter().map(|a| a.to_string()).collect::<Vec<_>>()
|
|
})))
|
|
}
|
|
|
|
#[tracing::instrument(name = "Fetching identifier", skip(repo))]
|
|
async fn identifier<S: Store>(
|
|
web::Query(alias_query): web::Query<AliasQuery>,
|
|
repo: web::Data<ArcRepo>,
|
|
) -> Result<HttpResponse, Error> {
|
|
let alias = match alias_query {
|
|
AliasQuery::Alias { alias } => Serde::into_inner(alias),
|
|
AliasQuery::Proxy { proxy } => {
|
|
let Some(alias) = repo.related(proxy).await? else {
|
|
return Ok(HttpResponse::NotFound().finish());
|
|
};
|
|
alias
|
|
}
|
|
};
|
|
|
|
let Some(identifier) = repo.identifier_from_alias(&alias).await? else {
|
|
// Invalid alias
|
|
return Ok(HttpResponse::NotFound().json(serde_json::json!({
|
|
"msg": "No identifiers associated with provided alias"
|
|
})));
|
|
};
|
|
|
|
Ok(HttpResponse::Ok().json(&serde_json::json!({
|
|
"msg": "ok",
|
|
"identifier": identifier.as_ref(),
|
|
})))
|
|
}
|
|
|
|
#[tracing::instrument(skip(repo, store))]
|
|
async fn healthz<S: Store>(
|
|
repo: web::Data<ArcRepo>,
|
|
store: web::Data<S>,
|
|
) -> Result<HttpResponse, Error> {
|
|
repo.health_check().await?;
|
|
store.health_check().await?;
|
|
Ok(HttpResponse::Ok().finish())
|
|
}
|
|
|
|
fn transform_error(error: actix_form_data::Error) -> actix_web::Error {
|
|
let error: Error = error.into();
|
|
let error: actix_web::Error = error.into();
|
|
error
|
|
}
|
|
|
|
fn build_client(config: &Configuration) -> Result<ClientWithMiddleware, Error> {
|
|
let client = reqwest::Client::builder()
|
|
.user_agent("pict-rs v0.5.0-main")
|
|
.use_rustls_tls()
|
|
.pool_max_idle_per_host(config.client.pool_size)
|
|
.build()
|
|
.map_err(UploadError::BuildClient)?;
|
|
|
|
Ok(ClientBuilder::new(client)
|
|
.with(TracingMiddleware::default())
|
|
.build())
|
|
}
|
|
|
|
fn configure_endpoints<S: Store + 'static, F: Fn(&mut web::ServiceConfig)>(
|
|
config: &mut web::ServiceConfig,
|
|
repo: ArcRepo,
|
|
store: S,
|
|
configuration: Configuration,
|
|
client: ClientWithMiddleware,
|
|
extra_config: F,
|
|
) {
|
|
config
|
|
.app_data(web::Data::new(repo))
|
|
.app_data(web::Data::new(store))
|
|
.app_data(web::Data::new(client))
|
|
.app_data(web::Data::new(configuration.clone()))
|
|
.route("/healthz", web::get().to(healthz::<S>))
|
|
.service(
|
|
web::scope("/image")
|
|
.service(
|
|
web::resource("")
|
|
.guard(guard::Post())
|
|
.route(web::post().to(upload::<S>)),
|
|
)
|
|
.service(
|
|
web::scope("/backgrounded")
|
|
.service(
|
|
web::resource("")
|
|
.guard(guard::Post())
|
|
.route(web::post().to(upload_backgrounded::<S>)),
|
|
)
|
|
.service(web::resource("/claim").route(web::get().to(claim_upload::<S>))),
|
|
)
|
|
.service(web::resource("/download").route(web::get().to(download::<S>)))
|
|
.service(
|
|
web::resource("/delete/{delete_token}/{filename}")
|
|
.route(web::delete().to(delete))
|
|
.route(web::get().to(delete)),
|
|
)
|
|
.service(
|
|
web::scope("/original")
|
|
.service(
|
|
web::resource("")
|
|
.route(web::get().to(serve_query::<S>))
|
|
.route(web::head().to(serve_query_head::<S>)),
|
|
)
|
|
.service(
|
|
web::resource("/{filename}")
|
|
.route(web::get().to(serve::<S>))
|
|
.route(web::head().to(serve_head::<S>)),
|
|
),
|
|
)
|
|
.service(
|
|
web::resource("/process.{ext}")
|
|
.route(web::get().to(process::<S>))
|
|
.route(web::head().to(process_head::<S>)),
|
|
)
|
|
.service(
|
|
web::resource("/process_backgrounded.{ext}")
|
|
.route(web::get().to(process_backgrounded::<S>)),
|
|
)
|
|
.service(
|
|
web::scope("/details")
|
|
.service(
|
|
web::scope("/original")
|
|
.service(web::resource("").route(web::get().to(details_query::<S>)))
|
|
.service(
|
|
web::resource("/{filename}").route(web::get().to(details::<S>)),
|
|
),
|
|
)
|
|
.service(
|
|
web::resource("/process.{ext}")
|
|
.route(web::get().to(process_details::<S>)),
|
|
),
|
|
),
|
|
)
|
|
.service(
|
|
web::scope("/internal")
|
|
.wrap(Internal(
|
|
configuration.server.api_key.as_ref().map(|s| s.to_owned()),
|
|
))
|
|
.service(web::resource("/import").route(web::post().to(import::<S>)))
|
|
.service(web::resource("/variants").route(web::delete().to(clean_variants)))
|
|
.service(web::resource("/purge").route(web::post().to(purge)))
|
|
.service(web::resource("/delete").route(web::post().to(delete_alias)))
|
|
.service(web::resource("/aliases").route(web::get().to(aliases)))
|
|
.service(web::resource("/identifier").route(web::get().to(identifier::<S>)))
|
|
.service(web::resource("/set_not_found").route(web::post().to(set_not_found)))
|
|
.service(web::resource("/hashes").route(web::get().to(page)))
|
|
.configure(extra_config),
|
|
);
|
|
}
|
|
|
|
fn spawn_cleanup(repo: ArcRepo, config: &Configuration) {
|
|
if config.server.read_only {
|
|
return;
|
|
}
|
|
|
|
crate::sync::spawn(async move {
|
|
let mut interval = actix_web::rt::time::interval(Duration::from_secs(30));
|
|
|
|
loop {
|
|
interval.tick().await;
|
|
|
|
if let Err(e) = queue::cleanup_outdated_variants(&repo).await {
|
|
tracing::warn!(
|
|
"Failed to spawn cleanup for outdated variants:{}",
|
|
format!("\n{e}\n{e:?}")
|
|
);
|
|
}
|
|
|
|
if let Err(e) = queue::cleanup_outdated_proxies(&repo).await {
|
|
tracing::warn!(
|
|
"Failed to spawn cleanup for outdated proxies:{}",
|
|
format!("\n{e}\n{e:?}")
|
|
);
|
|
}
|
|
}
|
|
});
|
|
}
|
|
|
|
fn spawn_workers<S>(
|
|
repo: ArcRepo,
|
|
store: S,
|
|
client: ClientWithMiddleware,
|
|
config: Configuration,
|
|
process_map: ProcessMap,
|
|
) where
|
|
S: Store + 'static,
|
|
{
|
|
crate::sync::spawn(queue::process_cleanup(
|
|
repo.clone(),
|
|
store.clone(),
|
|
config.clone(),
|
|
));
|
|
crate::sync::spawn(queue::process_images(
|
|
repo,
|
|
store,
|
|
client,
|
|
process_map,
|
|
config,
|
|
));
|
|
}
|
|
|
|
async fn launch_file_store<F: Fn(&mut web::ServiceConfig) + Send + Clone + 'static>(
|
|
repo: ArcRepo,
|
|
store: FileStore,
|
|
client: ClientWithMiddleware,
|
|
config: Configuration,
|
|
extra_config: F,
|
|
) -> std::io::Result<()> {
|
|
let process_map = ProcessMap::new();
|
|
|
|
let address = config.server.address;
|
|
|
|
spawn_cleanup(repo.clone(), &config);
|
|
|
|
HttpServer::new(move || {
|
|
let client = client.clone();
|
|
let store = store.clone();
|
|
let repo = repo.clone();
|
|
let config = config.clone();
|
|
let extra_config = extra_config.clone();
|
|
|
|
spawn_workers(
|
|
repo.clone(),
|
|
store.clone(),
|
|
client.clone(),
|
|
config.clone(),
|
|
process_map.clone(),
|
|
);
|
|
|
|
App::new()
|
|
.wrap(TracingLogger::default())
|
|
.wrap(Deadline)
|
|
.wrap(Metrics)
|
|
.wrap(Payload::new())
|
|
.app_data(web::Data::new(process_map.clone()))
|
|
.configure(move |sc| configure_endpoints(sc, repo, store, config, client, extra_config))
|
|
})
|
|
.bind(address)?
|
|
.run()
|
|
.await
|
|
}
|
|
|
|
async fn launch_object_store<F: Fn(&mut web::ServiceConfig) + Send + Clone + 'static>(
|
|
repo: ArcRepo,
|
|
store: ObjectStore,
|
|
client: ClientWithMiddleware,
|
|
config: Configuration,
|
|
extra_config: F,
|
|
) -> std::io::Result<()> {
|
|
let process_map = ProcessMap::new();
|
|
|
|
let address = config.server.address;
|
|
|
|
spawn_cleanup(repo.clone(), &config);
|
|
|
|
HttpServer::new(move || {
|
|
let client = client.clone();
|
|
let store = store.clone();
|
|
let repo = repo.clone();
|
|
let config = config.clone();
|
|
let extra_config = extra_config.clone();
|
|
|
|
spawn_workers(
|
|
repo.clone(),
|
|
store.clone(),
|
|
client.clone(),
|
|
config.clone(),
|
|
process_map.clone(),
|
|
);
|
|
|
|
App::new()
|
|
.wrap(TracingLogger::default())
|
|
.wrap(Deadline)
|
|
.wrap(Metrics)
|
|
.wrap(Payload::new())
|
|
.app_data(web::Data::new(process_map.clone()))
|
|
.configure(move |sc| configure_endpoints(sc, repo, store, config, client, extra_config))
|
|
})
|
|
.bind(address)?
|
|
.run()
|
|
.await
|
|
}
|
|
|
|
async fn migrate_inner<S1>(
|
|
repo: ArcRepo,
|
|
client: ClientWithMiddleware,
|
|
from: S1,
|
|
to: config::primitives::Store,
|
|
skip_missing_files: bool,
|
|
timeout: u64,
|
|
) -> color_eyre::Result<()>
|
|
where
|
|
S1: Store + 'static,
|
|
{
|
|
match to {
|
|
config::primitives::Store::Filesystem(config::Filesystem { path }) => {
|
|
let to = FileStore::build(path.clone(), repo.clone()).await?;
|
|
|
|
migrate_store(repo, from, to, skip_missing_files, timeout).await?
|
|
}
|
|
config::primitives::Store::ObjectStorage(config::primitives::ObjectStorage {
|
|
endpoint,
|
|
bucket_name,
|
|
use_path_style,
|
|
region,
|
|
access_key,
|
|
secret_key,
|
|
session_token,
|
|
signature_duration,
|
|
client_timeout,
|
|
public_endpoint,
|
|
}) => {
|
|
let to = ObjectStore::build(
|
|
endpoint.clone(),
|
|
bucket_name,
|
|
if use_path_style {
|
|
UrlStyle::Path
|
|
} else {
|
|
UrlStyle::VirtualHost
|
|
},
|
|
region,
|
|
access_key,
|
|
secret_key,
|
|
session_token,
|
|
signature_duration.unwrap_or(15),
|
|
client_timeout.unwrap_or(30),
|
|
public_endpoint,
|
|
repo.clone(),
|
|
)
|
|
.await?
|
|
.build(client);
|
|
|
|
migrate_store(repo, from, to, skip_missing_files, timeout).await?
|
|
}
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
impl<P: AsRef<Path>, T: serde::Serialize> ConfigSource<P, T> {
|
|
/// Initialize the pict-rs configuration
|
|
///
|
|
/// This takes an optional config_file path which is a valid pict-rs configuration file, and an
|
|
/// optional save_to path, which the generated configuration will be saved into. Since many
|
|
/// parameters have defaults, it can be useful to dump a valid configuration with default values to
|
|
/// see what is available for tweaking.
|
|
///
|
|
/// When running pict-rs as a library, configuration is limited to environment variables and
|
|
/// configuration files. Commandline options are not available.
|
|
///
|
|
/// ```rust
|
|
/// fn main() -> Result<(), Box<dyn std::error::Error>> {
|
|
/// pict_rs::ConfigSource::memory(serde_json::json!({
|
|
/// "server": {
|
|
/// "address": "127.0.0.1:8080"
|
|
/// },
|
|
/// "old_db": {
|
|
/// "path": "./old"
|
|
/// },
|
|
/// "repo": {
|
|
/// "type": "sled",
|
|
/// "path": "./sled-repo"
|
|
/// },
|
|
/// "store": {
|
|
/// "type": "filesystem",
|
|
/// "path": "./files"
|
|
/// }
|
|
/// })).init::<&str>(None)?;
|
|
///
|
|
/// Ok(())
|
|
/// }
|
|
/// ```
|
|
pub fn init<Q: AsRef<Path>>(
|
|
self,
|
|
save_to: Option<Q>,
|
|
) -> color_eyre::Result<PictRsConfiguration> {
|
|
config::configure_without_clap(self, save_to)
|
|
}
|
|
}
|
|
|
|
async fn export_handler(repo: web::Data<SledRepo>) -> Result<HttpResponse, Error> {
|
|
repo.export().await?;
|
|
|
|
Ok(HttpResponse::Created().json(&serde_json::json!({
|
|
"msg": "ok"
|
|
})))
|
|
}
|
|
|
|
fn sled_extra_config(sc: &mut web::ServiceConfig, repo: SledRepo) {
|
|
sc.app_data(web::Data::new(repo))
|
|
.service(web::resource("/export").route(web::post().to(export_handler)));
|
|
}
|
|
|
|
impl PictRsConfiguration {
|
|
/// Build the pict-rs configuration from commandline arguments
|
|
///
|
|
/// This is probably not useful for 3rd party applications that handle their own commandline
|
|
pub fn build_default() -> color_eyre::Result<Self> {
|
|
config::configure()
|
|
}
|
|
|
|
/// Install the default pict-rs tracer
|
|
///
|
|
/// This is probably not useful for 3rd party applications that install their own tracing
|
|
/// subscribers.
|
|
pub fn install_tracing(self) -> color_eyre::Result<Self> {
|
|
init_tracing(&self.config.tracing)?;
|
|
Ok(self)
|
|
}
|
|
|
|
pub fn install_metrics(self) -> color_eyre::Result<Self> {
|
|
if let Some(addr) = self.config.metrics.prometheus_address {
|
|
PrometheusBuilder::new()
|
|
.with_http_listener(addr)
|
|
.install()?;
|
|
}
|
|
|
|
Ok(self)
|
|
}
|
|
|
|
/// Run the pict-rs application
|
|
///
|
|
/// This must be called after `init_config`, or else the default configuration builder will run and
|
|
/// fail.
|
|
pub async fn run(self) -> color_eyre::Result<()> {
|
|
let PictRsConfiguration { config, operation } = self;
|
|
|
|
let client = build_client(&config)?;
|
|
|
|
match operation {
|
|
Operation::Run => (),
|
|
Operation::MigrateStore {
|
|
skip_missing_files,
|
|
from,
|
|
to,
|
|
} => {
|
|
let repo = Repo::open(config.repo.clone()).await?.to_arc();
|
|
|
|
match from {
|
|
config::primitives::Store::Filesystem(config::Filesystem { path }) => {
|
|
let from = FileStore::build(path.clone(), repo.clone()).await?;
|
|
migrate_inner(
|
|
repo,
|
|
client,
|
|
from,
|
|
to,
|
|
skip_missing_files,
|
|
config.media.process_timeout,
|
|
)
|
|
.await?;
|
|
}
|
|
config::primitives::Store::ObjectStorage(
|
|
config::primitives::ObjectStorage {
|
|
endpoint,
|
|
bucket_name,
|
|
use_path_style,
|
|
region,
|
|
access_key,
|
|
secret_key,
|
|
session_token,
|
|
signature_duration,
|
|
client_timeout,
|
|
public_endpoint,
|
|
},
|
|
) => {
|
|
let from = ObjectStore::build(
|
|
endpoint,
|
|
bucket_name,
|
|
if use_path_style {
|
|
UrlStyle::Path
|
|
} else {
|
|
UrlStyle::VirtualHost
|
|
},
|
|
region,
|
|
access_key,
|
|
secret_key,
|
|
session_token,
|
|
signature_duration.unwrap_or(15),
|
|
client_timeout.unwrap_or(30),
|
|
public_endpoint,
|
|
repo.clone(),
|
|
)
|
|
.await?
|
|
.build(client.clone());
|
|
|
|
migrate_inner(
|
|
repo,
|
|
client,
|
|
from,
|
|
to,
|
|
skip_missing_files,
|
|
config.media.process_timeout,
|
|
)
|
|
.await?;
|
|
}
|
|
}
|
|
|
|
return Ok(());
|
|
}
|
|
Operation::MigrateRepo { from, to } => {
|
|
let from = Repo::open(from).await?.to_arc();
|
|
let to = Repo::open(to).await?.to_arc();
|
|
|
|
repo::migrate_repo(from, to).await?;
|
|
return Ok(());
|
|
}
|
|
}
|
|
|
|
let repo = Repo::open(config.repo.clone()).await?;
|
|
|
|
if config.server.read_only {
|
|
tracing::warn!("Launching in READ ONLY mode");
|
|
}
|
|
|
|
match config.store.clone() {
|
|
config::Store::Filesystem(config::Filesystem { path }) => {
|
|
let arc_repo = repo.to_arc();
|
|
|
|
let store = FileStore::build(path, arc_repo.clone()).await?;
|
|
|
|
if arc_repo.get("migrate-0.4").await?.is_none() {
|
|
if let Some(old_repo) = repo_04::open(&config.old_repo)? {
|
|
repo::migrate_04(old_repo, arc_repo.clone(), store.clone(), config.clone())
|
|
.await?;
|
|
arc_repo
|
|
.set("migrate-0.4", Arc::from(b"migrated".to_vec()))
|
|
.await?;
|
|
}
|
|
}
|
|
|
|
match repo {
|
|
Repo::Sled(sled_repo) => {
|
|
launch_file_store(arc_repo, store, client, config, move |sc| {
|
|
sled_extra_config(sc, sled_repo.clone())
|
|
})
|
|
.await?;
|
|
}
|
|
Repo::Postgres(_) => {
|
|
launch_file_store(arc_repo, store, client, config, |_| {}).await?;
|
|
}
|
|
}
|
|
}
|
|
config::Store::ObjectStorage(config::ObjectStorage {
|
|
endpoint,
|
|
bucket_name,
|
|
use_path_style,
|
|
region,
|
|
access_key,
|
|
secret_key,
|
|
session_token,
|
|
signature_duration,
|
|
client_timeout,
|
|
public_endpoint,
|
|
}) => {
|
|
let arc_repo = repo.to_arc();
|
|
|
|
let store = ObjectStore::build(
|
|
endpoint,
|
|
bucket_name,
|
|
if use_path_style {
|
|
UrlStyle::Path
|
|
} else {
|
|
UrlStyle::VirtualHost
|
|
},
|
|
region,
|
|
access_key,
|
|
secret_key,
|
|
session_token,
|
|
signature_duration,
|
|
client_timeout,
|
|
public_endpoint,
|
|
arc_repo.clone(),
|
|
)
|
|
.await?
|
|
.build(client.clone());
|
|
|
|
if arc_repo.get("migrate-0.4").await?.is_none() {
|
|
if let Some(old_repo) = repo_04::open(&config.old_repo)? {
|
|
repo::migrate_04(old_repo, arc_repo.clone(), store.clone(), config.clone())
|
|
.await?;
|
|
arc_repo
|
|
.set("migrate-0.4", Arc::from(b"migrated".to_vec()))
|
|
.await?;
|
|
}
|
|
}
|
|
|
|
match repo {
|
|
Repo::Sled(sled_repo) => {
|
|
launch_object_store(arc_repo, store, client, config, move |sc| {
|
|
sled_extra_config(sc, sled_repo.clone())
|
|
})
|
|
.await?;
|
|
}
|
|
Repo::Postgres(_) => {
|
|
launch_object_store(arc_repo, store, client, config, |_| {}).await?;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
self::tmp_file::remove_tmp_dir().await?;
|
|
|
|
Ok(())
|
|
}
|
|
}
|
|
|
|
#[cfg(test)]
|
|
mod tests {
|
|
#[test]
|
|
fn source() {
|
|
let query = super::ProcessQuery {
|
|
source: super::ProcessSource::Source {
|
|
src: super::Serde::new(super::Alias::from_existing("example.png")),
|
|
},
|
|
operations: vec![("resize".into(), "200".into())],
|
|
};
|
|
let encoded = serde_urlencoded::to_string(&query).expect("Encoded");
|
|
let new_query: super::ProcessQuery = serde_urlencoded::from_str(&encoded).expect("Decoded");
|
|
// Don't compare entire query - "src" gets deserialized twice
|
|
assert_eq!(new_query.source, query.source);
|
|
|
|
assert!(new_query
|
|
.operations
|
|
.contains(&("resize".into(), "200".into())));
|
|
}
|
|
|
|
#[test]
|
|
fn alias() {
|
|
let query = super::ProcessQuery {
|
|
source: super::ProcessSource::Alias {
|
|
alias: super::Serde::new(super::Alias::from_existing("example.png")),
|
|
},
|
|
operations: vec![("resize".into(), "200".into())],
|
|
};
|
|
let encoded = serde_urlencoded::to_string(&query).expect("Encoded");
|
|
let new_query: super::ProcessQuery = serde_urlencoded::from_str(&encoded).expect("Decoded");
|
|
// Don't compare entire query - "alias" gets deserialized twice
|
|
assert_eq!(new_query.source, query.source);
|
|
|
|
assert!(new_query
|
|
.operations
|
|
.contains(&("resize".into(), "200".into())));
|
|
}
|
|
|
|
#[test]
|
|
fn url() {
|
|
let query = super::ProcessQuery {
|
|
source: super::ProcessSource::Proxy {
|
|
proxy: "http://example.com/image.png".parse().expect("valid url"),
|
|
},
|
|
operations: vec![("resize".into(), "200".into())],
|
|
};
|
|
let encoded = serde_urlencoded::to_string(&query).expect("Encoded");
|
|
let new_query: super::ProcessQuery = serde_urlencoded::from_str(&encoded).expect("Decoded");
|
|
// Don't compare entire query - "proxy" gets deserialized twice
|
|
assert_eq!(new_query.source, query.source);
|
|
|
|
assert!(new_query
|
|
.operations
|
|
.contains(&("resize".into(), "200".into())));
|
|
}
|
|
}
|