mod backgrounded; mod bytes_stream; mod concurrent_processor; mod config; mod details; mod discover; mod either; mod error; mod error_code; mod exiftool; mod ffmpeg; mod file; mod formats; mod future; mod generate; mod ingest; mod init_tracing; mod magick; mod middleware; mod migrate_store; mod process; mod processor; mod queue; mod range; mod read; mod repo; mod repo_04; mod serde_str; mod store; mod stream; mod sync; mod tmp_file; mod validate; use actix_form_data::{Field, Form, FormData, Multipart, Value}; use actix_web::{ guard, http::header::{CacheControl, CacheDirective, LastModified, Range, ACCEPT_RANGES}, web, App, HttpRequest, HttpResponse, HttpResponseBuilder, HttpServer, }; use details::{ApiDetails, HumanDate}; use future::WithTimeout; use futures_core::Stream; use metrics_exporter_prometheus::PrometheusBuilder; use middleware::{Metrics, Payload}; use repo::ArcRepo; use reqwest_middleware::{ClientBuilder, ClientWithMiddleware}; use reqwest_tracing::TracingMiddleware; use rusty_s3::UrlStyle; use std::{ marker::PhantomData, path::Path, path::PathBuf, sync::{Arc, OnceLock}, time::{Duration, SystemTime}, }; use streem::IntoStreamer; use tmp_file::{ArcTmpDir, TmpDir}; use tokio::sync::Semaphore; use tracing::Instrument; use tracing_actix_web::TracingLogger; use self::{ backgrounded::Backgrounded, concurrent_processor::ProcessMap, config::{Configuration, Operation}, details::Details, either::Either, error::{Error, UploadError}, formats::InputProcessableFormat, ingest::Session, init_tracing::init_tracing, middleware::{Deadline, Internal}, migrate_store::migrate_store, queue::queue_generate, repo::{sled::SledRepo, Alias, DeleteToken, Hash, Repo, UploadId, UploadResult}, serde_str::Serde, store::{file_store::FileStore, object_store::ObjectStore, Store}, stream::{empty, once}, }; pub use self::config::{ConfigSource, PictRsConfiguration}; const MEGABYTES: usize = 1024 * 1024; const MINUTES: u32 = 60; const HOURS: u32 = 60 * MINUTES; const DAYS: u32 = 24 * HOURS; const NOT_FOUND_KEY: &str = "404-alias"; static PROCESS_SEMAPHORE: OnceLock = OnceLock::new(); fn process_semaphore() -> &'static Semaphore { PROCESS_SEMAPHORE.get_or_init(|| { let permits = std::thread::available_parallelism() .map(usize::from) .unwrap_or(1) .saturating_sub(1) .max(1); crate::sync::bare_semaphore(permits) }) } async fn ensure_details( tmp_dir: &TmpDir, repo: &ArcRepo, store: &S, config: &Configuration, alias: &Alias, ) -> Result { let Some(identifier) = repo.identifier_from_alias(alias).await? else { return Err(UploadError::MissingAlias.into()); }; ensure_details_identifier(tmp_dir, repo, store, config, &identifier).await } async fn ensure_details_identifier( tmp_dir: &TmpDir, repo: &ArcRepo, store: &S, config: &Configuration, identifier: &Arc, ) -> Result { let details = repo.details(identifier).await?; if let Some(details) = details { tracing::debug!("details exist"); Ok(details) } else { if config.server.read_only { return Err(UploadError::ReadOnly.into()); } else if config.server.danger_dummy_mode { return Ok(Details::danger_dummy(formats::InternalFormat::Image( formats::ImageFormat::Png, ))); } tracing::debug!("generating new details from {:?}", identifier); let bytes_stream = store.to_bytes(identifier, None, None).await?; let new_details = Details::from_bytes( tmp_dir, config.media.process_timeout, bytes_stream.into_bytes(), ) .await?; tracing::debug!("storing details for {:?}", identifier); repo.relate_details(identifier, &new_details).await?; tracing::debug!("stored"); Ok(new_details) } } struct Upload(Value, PhantomData); impl FormData for Upload { type Item = Session; type Error = Error; fn form(req: &HttpRequest) -> Form { let tmp_dir = req .app_data::>() .expect("No TmpDir in request") .clone(); let repo = req .app_data::>() .expect("No repo in request") .clone(); let store = req .app_data::>() .expect("No store in request") .clone(); let client = req .app_data::>() .expect("No client in request") .clone(); let config = req .app_data::>() .expect("No configuration in request") .clone(); // Create a new Multipart Form validator // // This form is expecting a single array field, 'images' with at most 10 files in it Form::new() .max_files(config.server.max_file_count) .max_file_size(config.media.max_file_size * MEGABYTES) .transform_error(transform_error) .field( "images", Field::array(Field::file(move |filename, _, stream| { let tmp_dir = tmp_dir.clone(); let repo = repo.clone(); let store = store.clone(); let client = client.clone(); let config = config.clone(); metrics::counter!("pict-rs.files", "upload" => "inline").increment(1); let span = tracing::info_span!("file-upload", ?filename); Box::pin( async move { if config.server.read_only { return Err(UploadError::ReadOnly.into()); } let stream = crate::stream::from_err(stream); ingest::ingest( &tmp_dir, &repo, &**store, &client, stream, None, &config, ) .await } .instrument(span), ) })), ) } fn extract(value: Value) -> Result { Ok(Upload(value, PhantomData)) } } struct Import(Value, PhantomData); impl FormData for Import { type Item = Session; type Error = Error; fn form(req: &actix_web::HttpRequest) -> Form { let tmp_dir = req .app_data::>() .expect("No TmpDir in request") .clone(); let repo = req .app_data::>() .expect("No repo in request") .clone(); let store = req .app_data::>() .expect("No store in request") .clone(); let client = req .app_data::() .expect("No client in request") .clone(); let config = req .app_data::>() .expect("No configuration in request") .clone(); // Create a new Multipart Form validator for internal imports // // This form is expecting a single array field, 'images' with at most 10 files in it Form::new() .max_files(config.server.max_file_count) .max_file_size(config.media.max_file_size * MEGABYTES) .transform_error(transform_error) .field( "images", Field::array(Field::file(move |filename, _, stream| { let tmp_dir = tmp_dir.clone(); let repo = repo.clone(); let store = store.clone(); let client = client.clone(); let config = config.clone(); metrics::counter!("pict-rs.files", "import" => "inline").increment(1); let span = tracing::info_span!("file-import", ?filename); Box::pin( async move { if config.server.read_only { return Err(UploadError::ReadOnly.into()); } let stream = crate::stream::from_err(stream); ingest::ingest( &tmp_dir, &repo, &**store, &client, stream, Some(Alias::from_existing(&filename)), &config, ) .await } .instrument(span), ) })), ) } fn extract(value: Value) -> Result where Self: Sized, { Ok(Import(value, PhantomData)) } } /// Handle responding to successful uploads #[tracing::instrument(name = "Uploaded files", skip(value, repo, store, config))] async fn upload( Multipart(Upload(value, _)): Multipart>, tmp_dir: web::Data, repo: web::Data, store: web::Data, config: web::Data, ) -> Result { handle_upload(value, tmp_dir, repo, store, config).await } /// Handle responding to successful uploads #[tracing::instrument(name = "Imported files", skip(value, repo, store, config))] async fn import( Multipart(Import(value, _)): Multipart>, tmp_dir: web::Data, repo: web::Data, store: web::Data, config: web::Data, ) -> Result { handle_upload(value, tmp_dir, repo, store, config).await } /// Handle responding to successful uploads #[tracing::instrument(name = "Uploaded files", skip(value, repo, store, config))] async fn handle_upload( value: Value, tmp_dir: web::Data, repo: web::Data, store: web::Data, config: web::Data, ) -> Result { let images = value .map() .and_then(|mut m| m.remove("images")) .and_then(|images| images.array()) .ok_or(UploadError::NoFiles)?; let mut files = Vec::new(); let images = images .into_iter() .filter_map(|i| i.file()) .collect::>(); for image in &images { if let Some(alias) = image.result.alias() { tracing::debug!("Uploaded {} as {:?}", image.filename, alias); let delete_token = image.result.delete_token(); let details = ensure_details(&tmp_dir, &repo, &store, &config, alias).await?; files.push(serde_json::json!({ "file": alias.to_string(), "delete_token": delete_token.to_string(), "details": details.into_api_details(), })); } } for image in images { image.result.disarm(); } Ok(HttpResponse::Created().json(&serde_json::json!({ "msg": "ok", "files": files }))) } struct BackgroundedUpload(Value, PhantomData); impl FormData for BackgroundedUpload { type Item = Backgrounded; type Error = Error; fn form(req: &actix_web::HttpRequest) -> Form { // Create a new Multipart Form validator for backgrounded uploads // // This form is expecting a single array field, 'images' with at most 10 files in it let repo = req .app_data::>() .expect("No repo in request") .clone(); let store = req .app_data::>() .expect("No store in request") .clone(); let config = req .app_data::>() .expect("No configuration in request") .clone(); let read_only = config.server.read_only; Form::new() .max_files(config.server.max_file_count) .max_file_size(config.media.max_file_size * MEGABYTES) .transform_error(transform_error) .field( "images", Field::array(Field::file(move |filename, _, stream| { let repo = (**repo).clone(); let store = (**store).clone(); metrics::counter!("pict-rs.files", "upload" => "background").increment(1); let span = tracing::info_span!("file-proxy", ?filename); Box::pin( async move { if read_only { return Err(UploadError::ReadOnly.into()); } let stream = crate::stream::from_err(stream); Backgrounded::proxy(repo, store, stream).await } .instrument(span), ) })), ) } fn extract(value: Value) -> Result where Self: Sized, { Ok(BackgroundedUpload(value, PhantomData)) } } #[tracing::instrument(name = "Uploaded files", skip(value, repo))] async fn upload_backgrounded( Multipart(BackgroundedUpload(value, _)): Multipart>, repo: web::Data, ) -> Result { let images = value .map() .and_then(|mut m| m.remove("images")) .and_then(|images| images.array()) .ok_or(UploadError::NoFiles)?; let mut files = Vec::new(); let images = images .into_iter() .filter_map(|i| i.file()) .collect::>(); for image in &images { let upload_id = image.result.upload_id().expect("Upload ID exists"); let identifier = image.result.identifier().expect("Identifier exists"); queue::queue_ingest(&repo, identifier, upload_id, None).await?; files.push(serde_json::json!({ "upload_id": upload_id.to_string(), })); } for image in images { image.result.disarm(); } Ok(HttpResponse::Accepted().json(&serde_json::json!({ "msg": "ok", "uploads": files }))) } #[derive(Debug, serde::Deserialize)] struct ClaimQuery { upload_id: Serde, } /// Claim a backgrounded upload #[tracing::instrument(name = "Waiting on upload", skip_all)] async fn claim_upload( tmp_dir: web::Data, repo: web::Data, store: web::Data, config: web::Data, query: web::Query, ) -> Result { let upload_id = Serde::into_inner(query.into_inner().upload_id); match repo .wait(upload_id) .with_timeout(Duration::from_secs(10)) .await { Ok(wait_res) => { let upload_result = wait_res?; repo.claim(upload_id).await?; metrics::counter!("pict-rs.background.upload.claim").increment(1); match upload_result { UploadResult::Success { alias, token } => { let details = ensure_details(&tmp_dir, &repo, &store, &config, &alias).await?; Ok(HttpResponse::Ok().json(&serde_json::json!({ "msg": "ok", "files": [{ "file": alias.to_string(), "delete_token": token.to_string(), "details": details.into_api_details(), }] }))) } UploadResult::Failure { message, code } => Ok(HttpResponse::UnprocessableEntity() .json(&serde_json::json!({ "msg": message, "code": code, }))), } } Err(_) => Ok(HttpResponse::NoContent().finish()), } } #[derive(Debug, serde::Deserialize)] struct UrlQuery { url: String, #[serde(default)] backgrounded: bool, } async fn ingest_inline( stream: impl Stream> + 'static, tmp_dir: &TmpDir, repo: &ArcRepo, store: &S, client: &ClientWithMiddleware, config: &Configuration, ) -> Result<(Alias, DeleteToken, Details), Error> { let session = ingest::ingest(tmp_dir, repo, store, client, stream, None, config).await?; let alias = session.alias().expect("alias should exist").to_owned(); let details = ensure_details(tmp_dir, repo, store, config, &alias).await?; let delete_token = session.disarm(); Ok((alias, delete_token, details)) } /// download an image from a URL #[tracing::instrument(name = "Downloading file", skip(client, repo, store, config))] async fn download( client: web::Data, tmp_dir: web::Data, repo: web::Data, store: web::Data, config: web::Data, query: web::Query, ) -> Result { let stream = download_stream(&client, &query.url, &config).await?; if query.backgrounded { do_download_backgrounded(stream, repo, store).await } else { do_download_inline(stream, &tmp_dir, repo, store, &client, config).await } } async fn download_stream( client: &ClientWithMiddleware, url: &str, config: &Configuration, ) -> Result> + 'static, Error> { if config.server.read_only { return Err(UploadError::ReadOnly.into()); } let res = client.get(url).send().await?; if !res.status().is_success() { return Err(UploadError::Download(res.status()).into()); } let stream = crate::stream::limit( config.media.max_file_size * MEGABYTES, crate::stream::from_err(res.bytes_stream()), ); Ok(stream) } #[tracing::instrument( name = "Downloading file inline", skip(stream, repo, store, client, config) )] async fn do_download_inline( stream: impl Stream> + 'static, tmp_dir: &TmpDir, repo: web::Data, store: web::Data, client: &ClientWithMiddleware, config: web::Data, ) -> Result { metrics::counter!("pict-rs.files", "download" => "inline").increment(1); let (alias, delete_token, details) = ingest_inline(stream, tmp_dir, &repo, &store, client, &config).await?; Ok(HttpResponse::Created().json(&serde_json::json!({ "msg": "ok", "files": [{ "file": alias.to_string(), "delete_token": delete_token.to_string(), "details": details.into_api_details(), }] }))) } #[tracing::instrument(name = "Downloading file in background", skip(stream, repo, store))] async fn do_download_backgrounded( stream: impl Stream> + 'static, repo: web::Data, store: web::Data, ) -> Result { metrics::counter!("pict-rs.files", "download" => "background").increment(1); let backgrounded = Backgrounded::proxy((**repo).clone(), (**store).clone(), stream).await?; let upload_id = backgrounded.upload_id().expect("Upload ID exists"); let identifier = backgrounded.identifier().expect("Identifier exists"); queue::queue_ingest(&repo, identifier, upload_id, None).await?; backgrounded.disarm(); Ok(HttpResponse::Accepted().json(&serde_json::json!({ "msg": "ok", "uploads": [{ "upload_id": upload_id.to_string(), }] }))) } #[derive(Debug, serde::Deserialize)] struct PageQuery { slug: Option, timestamp: Option, limit: Option, } #[derive(serde::Serialize)] struct PageJson { limit: usize, #[serde(skip_serializing_if = "Option::is_none")] current: Option, #[serde(skip_serializing_if = "Option::is_none")] prev: Option, #[serde(skip_serializing_if = "Option::is_none")] next: Option, hashes: Vec, } #[derive(serde::Serialize)] struct HashJson { hex: String, aliases: Vec, details: Option, } /// Get a page of hashes #[tracing::instrument(name = "Hash Page", skip(repo))] async fn page( repo: web::Data, web::Query(PageQuery { slug, timestamp, limit, }): web::Query, ) -> Result { let limit = limit.unwrap_or(20); let page = if let Some(timestamp) = timestamp { repo.hash_page_by_date(timestamp.timestamp, limit).await? } else { repo.hash_page(slug, limit).await? }; let mut hashes = Vec::with_capacity(page.hashes.len()); for hash in &page.hashes { let hex = hash.to_hex(); let aliases = repo .aliases_for_hash(hash.clone()) .await? .into_iter() .map(|a| a.to_string()) .collect(); let identifier = repo.identifier(hash.clone()).await?; let details = if let Some(identifier) = identifier { repo.details(&identifier) .await? .map(|d| d.into_api_details()) } else { None }; hashes.push(HashJson { hex, aliases, details, }); } let page = PageJson { limit: page.limit, current: page.current(), prev: page.prev(), next: page.next(), hashes, }; Ok(HttpResponse::Ok().json(serde_json::json!({ "msg": "ok", "page": page, }))) } /// Delete aliases and files #[tracing::instrument(name = "Deleting file", skip(repo, config))] async fn delete( repo: web::Data, config: web::Data, path_entries: web::Path<(String, String)>, ) -> Result { if config.server.read_only { return Err(UploadError::ReadOnly.into()); } let (token, alias) = path_entries.into_inner(); let token = DeleteToken::from_existing(&token); let alias = Alias::from_existing(&alias); // delete alias inline queue::cleanup::alias(&repo, alias, token).await?; Ok(HttpResponse::NoContent().finish()) } #[derive(Debug, serde::Deserialize, serde::Serialize, PartialEq, Eq, PartialOrd, Ord)] #[serde(untagged)] enum ProcessSource { Source { src: Serde }, Alias { alias: Serde }, Proxy { proxy: url::Url }, } #[derive(Debug, serde::Deserialize, serde::Serialize, PartialEq, Eq, PartialOrd, Ord)] struct ProcessQuery { #[serde(flatten)] source: ProcessSource, #[serde(with = "tuple_vec_map", flatten)] operations: Vec<(String, String)>, } fn prepare_process( config: &Configuration, operations: Vec<(String, String)>, ext: &str, ) -> Result<(InputProcessableFormat, PathBuf, Vec), Error> { let operations = operations .into_iter() .filter(|(k, _)| config.media.filters.contains(&k.to_lowercase())) .collect::>(); let format = ext .parse::() .map_err(|_| UploadError::UnsupportedProcessExtension)?; let (thumbnail_path, thumbnail_args) = self::processor::build_chain(&operations, &format.to_string())?; Ok((format, thumbnail_path, thumbnail_args)) } #[tracing::instrument(name = "Fetching derived details", skip(repo, config))] async fn process_details( web::Query(ProcessQuery { source, operations }): web::Query, ext: web::Path, repo: web::Data, config: web::Data, ) -> Result { let alias = match source { ProcessSource::Alias { alias } | ProcessSource::Source { src: alias } => { Serde::into_inner(alias) } ProcessSource::Proxy { proxy } => { let Some(alias) = repo.related(proxy).await? else { return Ok(HttpResponse::NotFound().json(&serde_json::json!({ "msg": "No images associated with provided proxy url" }))); }; alias } }; let (_, thumbnail_path, _) = prepare_process(&config, operations, ext.as_str())?; let Some(hash) = repo.hash(&alias).await? else { // Invalid alias return Ok(HttpResponse::NotFound().json(&serde_json::json!({ "msg": "No images associated with provided alias", }))); }; let thumbnail_string = thumbnail_path.to_string_lossy().to_string(); if !config.server.read_only { repo.accessed_variant(hash.clone(), thumbnail_string.clone()) .await?; } let identifier = repo .variant_identifier(hash, thumbnail_string) .await? .ok_or(UploadError::MissingAlias)?; let details = repo.details(&identifier).await?; let details = details.ok_or(UploadError::NoFiles)?; Ok(HttpResponse::Ok().json(&details.into_api_details())) } async fn not_found_hash(repo: &ArcRepo) -> Result, Error> { let Some(not_found) = repo.get(NOT_FOUND_KEY).await? else { return Ok(None); }; let Some(alias) = Alias::from_slice(not_found.as_ref()) else { tracing::warn!("Couldn't parse not-found alias"); return Ok(None); }; let Some(hash) = repo.hash(&alias).await? else { tracing::warn!("No hash found for not-found alias"); return Ok(None); }; Ok(Some((alias, hash))) } /// Process files #[allow(clippy::too_many_arguments)] #[tracing::instrument( name = "Serving processed image", skip(tmp_dir, repo, store, client, config, process_map) )] async fn process( range: Option>, web::Query(ProcessQuery { source, operations }): web::Query, ext: web::Path, tmp_dir: web::Data, repo: web::Data, store: web::Data, client: web::Data, config: web::Data, process_map: web::Data, ) -> Result { let alias = match source { ProcessSource::Alias { alias } | ProcessSource::Source { src: alias } => { Serde::into_inner(alias) } ProcessSource::Proxy { proxy } => { let alias = if let Some(alias) = repo.related(proxy.clone()).await? { alias } else if !config.server.read_only { let stream = download_stream(&client, proxy.as_str(), &config).await?; let (alias, _, _) = ingest_inline(stream, &tmp_dir, &repo, &store, &client, &config).await?; repo.relate_url(proxy, alias.clone()).await?; alias } else { return Err(UploadError::ReadOnly.into()); }; if !config.server.read_only { repo.accessed_alias(alias.clone()).await?; } alias } }; let (format, thumbnail_path, thumbnail_args) = prepare_process(&config, operations, ext.as_str())?; let path_string = thumbnail_path.to_string_lossy().to_string(); let (hash, alias, not_found) = if let Some(hash) = repo.hash(&alias).await? { (hash, alias, false) } else { let Some((alias, hash)) = not_found_hash(&repo).await? else { return Ok(HttpResponse::NotFound().finish()); }; (hash, alias, true) }; if !config.server.read_only { repo.accessed_variant(hash.clone(), path_string.clone()) .await?; } let identifier_opt = repo.variant_identifier(hash.clone(), path_string).await?; if let Some(identifier) = identifier_opt { let details = ensure_details_identifier(&tmp_dir, &repo, &store, &config, &identifier).await?; if let Some(public_url) = store.public_url(&identifier) { return Ok(HttpResponse::SeeOther() .insert_header((actix_web::http::header::LOCATION, public_url.as_str())) .finish()); } return ranged_file_resp(&store, identifier, range, details, not_found).await; } if config.server.read_only { return Err(UploadError::ReadOnly.into()); } let original_details = ensure_details(&tmp_dir, &repo, &store, &config, &alias).await?; let (details, bytes) = generate::generate( &tmp_dir, &repo, &store, &process_map, format, thumbnail_path, thumbnail_args, &original_details, &config, hash, ) .await?; let (builder, stream) = if let Some(web::Header(range_header)) = range { if let Some(range) = range::single_bytes_range(&range_header) { let len = bytes.len() as u64; if let Some(content_range) = range::to_content_range(range, len) { let mut builder = HttpResponse::PartialContent(); builder.insert_header(content_range); let stream = range::chop_bytes(range, bytes, len)?; (builder, Either::left(Either::left(stream))) } else { ( HttpResponse::RangeNotSatisfiable(), Either::left(Either::right(empty())), ) } } else { return Err(UploadError::Range.into()); } } else if not_found { (HttpResponse::NotFound(), Either::right(once(Ok(bytes)))) } else { (HttpResponse::Ok(), Either::right(once(Ok(bytes)))) }; Ok(srv_response( builder, stream, details.media_type(), 7 * DAYS, details.system_time(), )) } #[tracing::instrument( name = "Serving processed image headers", skip(tmp_dir, repo, store, config) )] async fn process_head( range: Option>, web::Query(ProcessQuery { source, operations }): web::Query, ext: web::Path, tmp_dir: web::Data, repo: web::Data, store: web::Data, config: web::Data, ) -> Result { let alias = match source { ProcessSource::Alias { alias } | ProcessSource::Source { src: alias } => { Serde::into_inner(alias) } ProcessSource::Proxy { proxy } => { let Some(alias) = repo.related(proxy).await? else { return Ok(HttpResponse::NotFound().finish()); }; alias } }; let (_, thumbnail_path, _) = prepare_process(&config, operations, ext.as_str())?; let path_string = thumbnail_path.to_string_lossy().to_string(); let Some(hash) = repo.hash(&alias).await? else { // Invalid alias return Ok(HttpResponse::NotFound().finish()); }; if !config.server.read_only { repo.accessed_variant(hash.clone(), path_string.clone()) .await?; } let identifier_opt = repo.variant_identifier(hash.clone(), path_string).await?; if let Some(identifier) = identifier_opt { let details = ensure_details_identifier(&tmp_dir, &repo, &store, &config, &identifier).await?; if let Some(public_url) = store.public_url(&identifier) { return Ok(HttpResponse::SeeOther() .insert_header((actix_web::http::header::LOCATION, public_url.as_str())) .finish()); } return ranged_file_head_resp(&store, identifier, range, details).await; } Ok(HttpResponse::NotFound().finish()) } /// Process files #[tracing::instrument(name = "Spawning image process", skip(repo))] async fn process_backgrounded( web::Query(ProcessQuery { source, operations }): web::Query, ext: web::Path, repo: web::Data, config: web::Data, ) -> Result { let source = match source { ProcessSource::Alias { alias } | ProcessSource::Source { src: alias } => { Serde::into_inner(alias) } ProcessSource::Proxy { proxy } => { let Some(alias) = repo.related(proxy).await? else { return Ok(HttpResponse::NotFound().finish()); }; alias } }; let (target_format, process_path, process_args) = prepare_process(&config, operations, ext.as_str())?; let path_string = process_path.to_string_lossy().to_string(); let Some(hash) = repo.hash(&source).await? else { // Invalid alias return Ok(HttpResponse::BadRequest().finish()); }; let identifier_opt = repo.variant_identifier(hash.clone(), path_string).await?; if identifier_opt.is_some() { return Ok(HttpResponse::Accepted().finish()); } if config.server.read_only { return Err(UploadError::ReadOnly.into()); } queue_generate(&repo, target_format, source, process_path, process_args).await?; Ok(HttpResponse::Accepted().finish()) } /// Fetch file details #[tracing::instrument(name = "Fetching query details", skip(repo, store, config))] async fn details_query( web::Query(alias_query): web::Query, tmp_dir: web::Data, repo: web::Data, store: web::Data, config: web::Data, ) -> Result { let alias = match alias_query { AliasQuery::Alias { alias } => Serde::into_inner(alias), AliasQuery::Proxy { proxy } => { let Some(alias) = repo.related(proxy).await? else { return Ok(HttpResponse::NotFound().json(&serde_json::json!({ "msg": "Provided proxy URL has not been cached", }))); }; alias } }; do_details(alias, tmp_dir, repo, store, config).await } /// Fetch file details #[tracing::instrument(name = "Fetching details", skip(tmp_dir, repo, store, config))] async fn details( alias: web::Path>, tmp_dir: web::Data, repo: web::Data, store: web::Data, config: web::Data, ) -> Result { do_details( Serde::into_inner(alias.into_inner()), tmp_dir, repo, store, config, ) .await } async fn do_details( alias: Alias, tmp_dir: web::Data, repo: web::Data, store: web::Data, config: web::Data, ) -> Result { let details = ensure_details(&tmp_dir, &repo, &store, &config, &alias).await?; Ok(HttpResponse::Ok().json(&details.into_api_details())) } /// Serve files based on alias query #[tracing::instrument(name = "Serving file query", skip(repo, store, client, config))] async fn serve_query( range: Option>, web::Query(alias_query): web::Query, tmp_dir: web::Data, repo: web::Data, store: web::Data, client: web::Data, config: web::Data, ) -> Result { let alias = match alias_query { AliasQuery::Alias { alias } => Serde::into_inner(alias), AliasQuery::Proxy { proxy } => { let alias = if let Some(alias) = repo.related(proxy.clone()).await? { alias } else if !config.server.read_only { let stream = download_stream(&client, proxy.as_str(), &config).await?; let (alias, _, _) = ingest_inline(stream, &tmp_dir, &repo, &store, &client, &config).await?; repo.relate_url(proxy, alias.clone()).await?; alias } else { return Err(UploadError::ReadOnly.into()); }; if !config.server.read_only { repo.accessed_alias(alias.clone()).await?; } alias } }; do_serve(range, alias, tmp_dir, repo, store, config).await } /// Serve files #[tracing::instrument(name = "Serving file", skip(repo, store, config))] async fn serve( range: Option>, alias: web::Path>, tmp_dir: web::Data, repo: web::Data, store: web::Data, config: web::Data, ) -> Result { do_serve( range, Serde::into_inner(alias.into_inner()), tmp_dir, repo, store, config, ) .await } async fn do_serve( range: Option>, alias: Alias, tmp_dir: web::Data, repo: web::Data, store: web::Data, config: web::Data, ) -> Result { let (hash, alias, not_found) = if let Some(hash) = repo.hash(&alias).await? { (hash, alias, false) } else { let Some((alias, hash)) = not_found_hash(&repo).await? else { return Ok(HttpResponse::NotFound().finish()); }; (hash, alias, true) }; let Some(identifier) = repo.identifier(hash.clone()).await? else { tracing::warn!("Original File identifier for hash {hash:?} is missing, queue cleanup task",); crate::queue::cleanup_hash(&repo, hash).await?; return Ok(HttpResponse::NotFound().finish()); }; let details = ensure_details(&tmp_dir, &repo, &store, &config, &alias).await?; if let Some(public_url) = store.public_url(&identifier) { return Ok(HttpResponse::SeeOther() .insert_header((actix_web::http::header::LOCATION, public_url.as_str())) .finish()); } ranged_file_resp(&store, identifier, range, details, not_found).await } #[tracing::instrument(name = "Serving query file headers", skip(repo, store, config))] async fn serve_query_head( range: Option>, web::Query(alias_query): web::Query, tmp_dir: web::Data, repo: web::Data, store: web::Data, config: web::Data, ) -> Result { let alias = match alias_query { AliasQuery::Alias { alias } => Serde::into_inner(alias), AliasQuery::Proxy { proxy } => { let Some(alias) = repo.related(proxy).await? else { return Ok(HttpResponse::NotFound().finish()); }; alias } }; do_serve_head(range, alias, tmp_dir, repo, store, config).await } #[tracing::instrument(name = "Serving file headers", skip(repo, store, config))] async fn serve_head( range: Option>, alias: web::Path>, tmp_dir: web::Data, repo: web::Data, store: web::Data, config: web::Data, ) -> Result { do_serve_head( range, Serde::into_inner(alias.into_inner()), tmp_dir, repo, store, config, ) .await } async fn do_serve_head( range: Option>, alias: Alias, tmp_dir: web::Data, repo: web::Data, store: web::Data, config: web::Data, ) -> Result { let Some(identifier) = repo.identifier_from_alias(&alias).await? else { // Invalid alias return Ok(HttpResponse::NotFound().finish()); }; let details = ensure_details(&tmp_dir, &repo, &store, &config, &alias).await?; if let Some(public_url) = store.public_url(&identifier) { return Ok(HttpResponse::SeeOther() .insert_header((actix_web::http::header::LOCATION, public_url.as_str())) .finish()); } ranged_file_head_resp(&store, identifier, range, details).await } async fn ranged_file_head_resp( store: &S, identifier: Arc, range: Option>, details: Details, ) -> Result { let builder = if let Some(web::Header(range_header)) = range { //Range header exists - return as ranged if let Some(range) = range::single_bytes_range(&range_header) { let len = store.len(&identifier).await?; if let Some(content_range) = range::to_content_range(range, len) { let mut builder = HttpResponse::PartialContent(); builder.insert_header(content_range); builder } else { HttpResponse::RangeNotSatisfiable() } } else { return Err(UploadError::Range.into()); } } else { // no range header HttpResponse::Ok() }; Ok(srv_head( builder, details.media_type(), 7 * DAYS, details.system_time(), ) .finish()) } async fn ranged_file_resp( store: &S, identifier: Arc, range: Option>, details: Details, not_found: bool, ) -> Result { let (builder, stream) = if let Some(web::Header(range_header)) = range { //Range header exists - return as ranged if let Some(range) = range::single_bytes_range(&range_header) { let len = store.len(&identifier).await?; if let Some(content_range) = range::to_content_range(range, len) { let mut builder = HttpResponse::PartialContent(); builder.insert_header(content_range); ( builder, Either::left(Either::left( range::chop_store(range, store, &identifier, len).await?, )), ) } else { ( HttpResponse::RangeNotSatisfiable(), Either::left(Either::right(empty())), ) } } else { return Err(UploadError::Range.into()); } } else { //No Range header in the request - return the entire document let stream = crate::stream::from_err(store.to_stream(&identifier, None, None).await?); if not_found { (HttpResponse::NotFound(), Either::right(stream)) } else { (HttpResponse::Ok(), Either::right(stream)) } }; Ok(srv_response( builder, stream, details.media_type(), 7 * DAYS, details.system_time(), )) } // A helper method to produce responses with proper cache headers fn srv_response( builder: HttpResponseBuilder, stream: S, ext: mime::Mime, expires: u32, modified: SystemTime, ) -> HttpResponse where S: Stream> + 'static, E: std::error::Error + 'static, actix_web::Error: From, { let stream = crate::stream::timeout(Duration::from_secs(5), stream); let stream = streem::try_from_fn(|yielder| async move { let stream = std::pin::pin!(stream); let mut streamer = stream.into_streamer(); while let Some(res) = streamer.next().await { tracing::trace!("srv_response: looping"); let item = res.map_err(Error::from)??; yielder.yield_ok(item).await; } Ok(()) as Result<(), actix_web::Error> }); srv_head(builder, ext, expires, modified).streaming(stream) } // A helper method to produce responses with proper cache headers fn srv_head( mut builder: HttpResponseBuilder, ext: mime::Mime, expires: u32, modified: SystemTime, ) -> HttpResponseBuilder { builder .insert_header(LastModified(modified.into())) .insert_header(CacheControl(vec![ CacheDirective::Public, CacheDirective::MaxAge(expires), CacheDirective::Extension("immutable".to_owned(), None), ])) .insert_header((ACCEPT_RANGES, "bytes")) .content_type(ext.to_string()); builder } #[derive(serde::Serialize)] struct PruneResponse { complete: bool, progress: u64, total: u64, } #[derive(Debug, serde::Deserialize)] struct PruneQuery { force: bool, } #[tracing::instrument(name = "Prune missing identifiers", skip(repo))] async fn prune_missing( repo: web::Data, query: Option>, ) -> Result { let total = repo.size().await?; let progress = if let Some(progress) = repo.get("prune-missing-queued").await? { progress .as_ref() .try_into() .map(u64::from_be_bytes) .unwrap_or(0) } else { 0 }; let complete = repo.get("prune-missing-complete").await?.is_some(); let started = repo.get("prune-missing-started").await?.is_some(); if !started || query.is_some_and(|q| q.force) { queue::prune_missing(&repo).await?; } Ok(HttpResponse::Ok().json(PruneResponse { complete, progress, total, })) } #[tracing::instrument(name = "Spawning variant cleanup", skip(repo, config))] async fn clean_variants( repo: web::Data, config: web::Data, ) -> Result { if config.server.read_only { return Err(UploadError::ReadOnly.into()); } queue::cleanup_all_variants(&repo).await?; Ok(HttpResponse::NoContent().finish()) } #[derive(Debug, serde::Deserialize)] #[serde(untagged)] enum AliasQuery { Proxy { proxy: url::Url }, Alias { alias: Serde }, } #[tracing::instrument(name = "Setting 404 Image", skip(repo, config))] async fn set_not_found( json: web::Json, repo: web::Data, client: web::Data, config: web::Data, ) -> Result { if config.server.read_only { return Err(UploadError::ReadOnly.into()); } let alias = match json.into_inner() { AliasQuery::Alias { alias } => Serde::into_inner(alias), AliasQuery::Proxy { .. } => { return Ok(HttpResponse::BadRequest().json(serde_json::json!({ "msg": "Cannot use proxied media as Not Found image", }))); } }; if repo.hash(&alias).await?.is_none() { return Ok(HttpResponse::BadRequest().json(serde_json::json!({ "msg": "No hash associated with provided alias" }))); } repo.set(NOT_FOUND_KEY, alias.to_bytes().into()).await?; Ok(HttpResponse::Created().json(serde_json::json!({ "msg": "ok", }))) } #[tracing::instrument(name = "Purging file", skip(repo, config))] async fn purge( web::Query(alias_query): web::Query, repo: web::Data, config: web::Data, ) -> Result { if config.server.read_only { return Err(UploadError::ReadOnly.into()); } let alias = match alias_query { AliasQuery::Alias { alias } => Serde::into_inner(alias), AliasQuery::Proxy { proxy } => { let Some(alias) = repo.related(proxy).await? else { return Ok(HttpResponse::NotFound().finish()); }; alias } }; let aliases = repo.aliases_from_alias(&alias).await?; let Some(hash) = repo.hash(&alias).await? else { return Ok(HttpResponse::BadRequest().json(&serde_json::json!({ "msg": "No images associated with provided alias", }))); }; queue::cleanup_hash(&repo, hash).await?; Ok(HttpResponse::Ok().json(&serde_json::json!({ "msg": "ok", "aliases": aliases.iter().map(|a| a.to_string()).collect::>() }))) } #[tracing::instrument(name = "Deleting alias", skip(repo, config))] async fn delete_alias( web::Query(alias_query): web::Query, repo: web::Data, config: web::Data, ) -> Result { if config.server.read_only { return Err(UploadError::ReadOnly.into()); } let alias = match alias_query { AliasQuery::Alias { alias } => Serde::into_inner(alias), AliasQuery::Proxy { proxy } => { let Some(alias) = repo.related(proxy).await? else { return Ok(HttpResponse::NotFound().finish()); }; alias } }; if let Some(token) = repo.delete_token(&alias).await? { queue::cleanup_alias(&repo, alias, token).await?; } else { return Ok(HttpResponse::NotFound().finish()); } Ok(HttpResponse::Ok().json(&serde_json::json!({ "msg": "ok", }))) } #[tracing::instrument(name = "Fetching aliases", skip(repo))] async fn aliases( web::Query(alias_query): web::Query, repo: web::Data, ) -> Result { let alias = match alias_query { AliasQuery::Alias { alias } => Serde::into_inner(alias), AliasQuery::Proxy { proxy } => { let Some(alias) = repo.related(proxy).await? else { return Ok(HttpResponse::NotFound().finish()); }; alias } }; let aliases = repo.aliases_from_alias(&alias).await?; Ok(HttpResponse::Ok().json(&serde_json::json!({ "msg": "ok", "aliases": aliases.iter().map(|a| a.to_string()).collect::>() }))) } #[tracing::instrument(name = "Fetching identifier", skip(repo))] async fn identifier( web::Query(alias_query): web::Query, repo: web::Data, ) -> Result { let alias = match alias_query { AliasQuery::Alias { alias } => Serde::into_inner(alias), AliasQuery::Proxy { proxy } => { let Some(alias) = repo.related(proxy).await? else { return Ok(HttpResponse::NotFound().finish()); }; alias } }; let Some(identifier) = repo.identifier_from_alias(&alias).await? else { // Invalid alias return Ok(HttpResponse::NotFound().json(serde_json::json!({ "msg": "No identifiers associated with provided alias" }))); }; Ok(HttpResponse::Ok().json(&serde_json::json!({ "msg": "ok", "identifier": identifier.as_ref(), }))) } #[tracing::instrument(skip(repo, store))] async fn healthz( repo: web::Data, store: web::Data, ) -> Result { repo.health_check().await?; store.health_check().await?; Ok(HttpResponse::Ok().finish()) } fn transform_error(error: actix_form_data::Error) -> actix_web::Error { let error: Error = error.into(); let error: actix_web::Error = error.into(); error } fn build_client() -> Result { let client = reqwest::Client::builder() .user_agent("pict-rs v0.5.0-main") .use_rustls_tls() .build() .map_err(UploadError::BuildClient)?; Ok(ClientBuilder::new(client) .with(TracingMiddleware::default()) .build()) } fn configure_endpoints( config: &mut web::ServiceConfig, repo: ArcRepo, store: S, configuration: Configuration, client: ClientWithMiddleware, extra_config: F, ) { config .app_data(web::Data::new(repo)) .app_data(web::Data::new(store)) .app_data(web::Data::new(client)) .app_data(web::Data::new(configuration.clone())) .route("/healthz", web::get().to(healthz::)) .service( web::scope("/image") .service( web::resource("") .guard(guard::Post()) .route(web::post().to(upload::)), ) .service( web::scope("/backgrounded") .service( web::resource("") .guard(guard::Post()) .route(web::post().to(upload_backgrounded::)), ) .service(web::resource("/claim").route(web::get().to(claim_upload::))), ) .service(web::resource("/download").route(web::get().to(download::))) .service( web::resource("/delete/{delete_token}/{filename}") .route(web::delete().to(delete)) .route(web::get().to(delete)), ) .service( web::scope("/original") .service( web::resource("") .route(web::get().to(serve_query::)) .route(web::head().to(serve_query_head::)), ) .service( web::resource("/{filename}") .route(web::get().to(serve::)) .route(web::head().to(serve_head::)), ), ) .service( web::resource("/process.{ext}") .route(web::get().to(process::)) .route(web::head().to(process_head::)), ) .service( web::resource("/process_backgrounded.{ext}") .route(web::get().to(process_backgrounded::)), ) .service( web::scope("/details") .service( web::scope("/original") .service(web::resource("").route(web::get().to(details_query::))) .service( web::resource("/{filename}").route(web::get().to(details::)), ), ) .service( web::resource("/process.{ext}") .route(web::get().to(process_details::)), ), ), ) .service( web::scope("/internal") .wrap(Internal( configuration.server.api_key.as_ref().map(|s| s.to_owned()), )) .service(web::resource("/import").route(web::post().to(import::))) .service(web::resource("/variants").route(web::delete().to(clean_variants))) .service(web::resource("/purge").route(web::post().to(purge))) .service(web::resource("/delete").route(web::post().to(delete_alias))) .service(web::resource("/aliases").route(web::get().to(aliases))) .service(web::resource("/identifier").route(web::get().to(identifier::))) .service(web::resource("/set_not_found").route(web::post().to(set_not_found))) .service(web::resource("/hashes").route(web::get().to(page))) .service(web::resource("/prune_missing").route(web::post().to(prune_missing))) .configure(extra_config), ); } fn spawn_cleanup(repo: ArcRepo, config: &Configuration) { if config.server.read_only { return; } crate::sync::spawn("queue-cleanup", async move { let mut interval = tokio::time::interval(Duration::from_secs(30)); loop { tracing::trace!("queue_cleanup: looping"); interval.tick().await; if let Err(e) = queue::cleanup_outdated_variants(&repo).await { tracing::warn!( "Failed to spawn cleanup for outdated variants:{}", format!("\n{e}\n{e:?}") ); } if let Err(e) = queue::cleanup_outdated_proxies(&repo).await { tracing::warn!( "Failed to spawn cleanup for outdated proxies:{}", format!("\n{e}\n{e:?}") ); } } }); } fn spawn_workers( tmp_dir: ArcTmpDir, repo: ArcRepo, store: S, client: ClientWithMiddleware, config: Configuration, process_map: ProcessMap, ) where S: Store + 'static, { crate::sync::spawn( "cleanup-worker", queue::process_cleanup(repo.clone(), store.clone(), config.clone()), ); crate::sync::spawn( "process-worker", queue::process_images(tmp_dir, repo, store, client, process_map, config), ); } async fn launch_file_store( tmp_dir: ArcTmpDir, repo: ArcRepo, store: FileStore, client: ClientWithMiddleware, config: Configuration, extra_config: F, ) -> std::io::Result<()> { let process_map = ProcessMap::new(); let address = config.server.address; spawn_cleanup(repo.clone(), &config); HttpServer::new(move || { let tmp_dir = tmp_dir.clone(); let client = client.clone(); let store = store.clone(); let repo = repo.clone(); let config = config.clone(); let extra_config = extra_config.clone(); spawn_workers( tmp_dir.clone(), repo.clone(), store.clone(), client.clone(), config.clone(), process_map.clone(), ); App::new() .wrap(TracingLogger::default()) .wrap(Deadline) .wrap(Metrics) .wrap(Payload::new()) .app_data(web::Data::new(process_map.clone())) .app_data(web::Data::new(tmp_dir)) .configure(move |sc| configure_endpoints(sc, repo, store, config, client, extra_config)) }) .bind(address)? .run() .await } async fn launch_object_store( tmp_dir: ArcTmpDir, repo: ArcRepo, store: ObjectStore, client: ClientWithMiddleware, config: Configuration, extra_config: F, ) -> std::io::Result<()> { let process_map = ProcessMap::new(); let address = config.server.address; spawn_cleanup(repo.clone(), &config); HttpServer::new(move || { let tmp_dir = tmp_dir.clone(); let client = client.clone(); let store = store.clone(); let repo = repo.clone(); let config = config.clone(); let extra_config = extra_config.clone(); spawn_workers( tmp_dir.clone(), repo.clone(), store.clone(), client.clone(), config.clone(), process_map.clone(), ); App::new() .wrap(TracingLogger::default()) .wrap(Deadline) .wrap(Metrics) .wrap(Payload::new()) .app_data(web::Data::new(process_map.clone())) .app_data(web::Data::new(tmp_dir)) .configure(move |sc| configure_endpoints(sc, repo, store, config, client, extra_config)) }) .bind(address)? .run() .await } #[allow(clippy::too_many_arguments)] async fn migrate_inner( tmp_dir: ArcTmpDir, repo: ArcRepo, client: ClientWithMiddleware, from: S1, to: config::primitives::Store, skip_missing_files: bool, timeout: u64, concurrency: usize, ) -> color_eyre::Result<()> where S1: Store + 'static, { match to { config::primitives::Store::Filesystem(config::Filesystem { path }) => { let to = FileStore::build(path.clone(), repo.clone()).await?; migrate_store( tmp_dir, repo, from, to, skip_missing_files, timeout, concurrency, ) .await? } config::primitives::Store::ObjectStorage(config::primitives::ObjectStorage { endpoint, bucket_name, use_path_style, region, access_key, secret_key, session_token, signature_duration, client_timeout, public_endpoint, }) => { let to = ObjectStore::build( endpoint.clone(), bucket_name, if use_path_style { UrlStyle::Path } else { UrlStyle::VirtualHost }, region, access_key, secret_key, session_token, signature_duration.unwrap_or(15), client_timeout.unwrap_or(30), public_endpoint, repo.clone(), ) .await? .build(client); migrate_store( tmp_dir, repo, from, to, skip_missing_files, timeout, concurrency, ) .await? } } Ok(()) } impl, T: serde::Serialize> ConfigSource { /// Initialize the pict-rs configuration /// /// This takes an optional config_file path which is a valid pict-rs configuration file, and an /// optional save_to path, which the generated configuration will be saved into. Since many /// parameters have defaults, it can be useful to dump a valid configuration with default values to /// see what is available for tweaking. /// /// When running pict-rs as a library, configuration is limited to environment variables and /// configuration files. Commandline options are not available. /// /// ```rust /// fn main() -> Result<(), Box> { /// let configuration = pict_rs::ConfigSource::memory(serde_json::json!({ /// "server": { /// "address": "127.0.0.1:8080" /// }, /// "repo": { /// "type": "sled", /// "path": "./sled-repo" /// }, /// "store": { /// "type": "filesystem", /// "path": "./files" /// } /// })).init::<&str>(None)?; /// /// Ok(()) /// } /// ``` pub fn init>( self, save_to: Option, ) -> color_eyre::Result { config::configure_without_clap(self, save_to) } } async fn export_handler(repo: web::Data) -> Result { repo.export().await?; Ok(HttpResponse::Created().json(&serde_json::json!({ "msg": "ok" }))) } fn sled_extra_config(sc: &mut web::ServiceConfig, repo: SledRepo) { sc.app_data(web::Data::new(repo)) .service(web::resource("/export").route(web::post().to(export_handler))); } impl PictRsConfiguration { /// Build the pict-rs configuration from commandline arguments /// /// This is probably not useful for 3rd party applications that handle their own commandline pub fn build_default() -> color_eyre::Result { config::configure() } /// Install the default pict-rs tracer /// /// This is probably not useful for 3rd party applications that install their own tracing /// subscribers. pub fn install_tracing(self) -> color_eyre::Result { init_tracing(&self.config.tracing)?; Ok(self) } /// Install the configured pict-rs metrics collector /// /// This is a no-op if pict-rs is not configured to export metrics. Applications that register /// their own metrics collectors shouldn't call this method. pub fn install_metrics(self) -> color_eyre::Result { if let Some(addr) = self.config.metrics.prometheus_address { PrometheusBuilder::new() .with_http_listener(addr) .install()?; tracing::info!("Starting prometheus endpoint on {addr}"); } Ok(self) } /// Run the pict-rs application on a tokio `LocalSet` /// /// This must be called from within `tokio::main` directly /// /// Example: /// ```rust /// #[tokio::main] /// async fn main() -> color_eyre::Result<()> { /// let pict_rs_server = pict_rs::ConfigSource::memory(serde_json::json!({ /// "repo": { /// "type": "sled", /// "path": "/tmp/pict-rs/run-on-localset/sled-repo", /// }, /// "store": { /// "type": "filesystem", /// "path": "/tmp/pict-rs/run-on-localset/files", /// }, /// })) /// .init::<&str>(None)? /// .run_on_localset(); /// /// let _ = tokio::time::timeout(std::time::Duration::from_secs(1), pict_rs_server).await; /// /// Ok(()) /// } /// ``` pub async fn run_on_localset(self) -> color_eyre::Result<()> { tokio::task::LocalSet::new().run_until(self.run()).await } /// Run the pict-rs application /// /// This must be called from within a tokio `LocalSet`, which is created by default for /// actix-rt runtimes, and by tokio_uring /// /// Example: /// ```rust,ignore /// fn main() -> color_eyre::Result<()> { /// tokio_uring::start(async move { /// let pict_rs_server = pict_rs::ConfigSource::memory(serde_json::json!({ /// "repo": { /// "type": "sled", /// "path": "/tmp/pict-rs/run/sled-repo", /// }, /// "store": { /// "type": "filesystem", /// "path": "/tmp/pict-rs/run/files", /// }, /// })) /// .init::<&str>(None)? /// .run(); /// /// let _ = tokio::time::timeout(std::time::Duration::from_secs(1), pict_rs_server).await; /// /// Ok(()) /// }) /// } /// ``` pub async fn run(self) -> color_eyre::Result<()> { let PictRsConfiguration { config, operation } = self; let tmp_dir = TmpDir::init(&config.server.temporary_directory).await?; let client = build_client()?; match operation { Operation::Run => (), Operation::MigrateStore { skip_missing_files, concurrency, from, to, } => { let repo = Repo::open(config.repo.clone()).await?.to_arc(); match from { config::primitives::Store::Filesystem(config::Filesystem { path }) => { let from = FileStore::build(path.clone(), repo.clone()).await?; migrate_inner( tmp_dir, repo, client, from, to, skip_missing_files, config.media.process_timeout, concurrency, ) .await?; } config::primitives::Store::ObjectStorage( config::primitives::ObjectStorage { endpoint, bucket_name, use_path_style, region, access_key, secret_key, session_token, signature_duration, client_timeout, public_endpoint, }, ) => { let from = ObjectStore::build( endpoint, bucket_name, if use_path_style { UrlStyle::Path } else { UrlStyle::VirtualHost }, region, access_key, secret_key, session_token, signature_duration.unwrap_or(15), client_timeout.unwrap_or(30), public_endpoint, repo.clone(), ) .await? .build(client.clone()); migrate_inner( tmp_dir, repo, client, from, to, skip_missing_files, config.media.process_timeout, concurrency, ) .await?; } } return Ok(()); } Operation::MigrateRepo { from, to } => { let from = Repo::open(from).await?.to_arc(); let to = Repo::open(to).await?.to_arc(); repo::migrate_repo(from, to).await?; return Ok(()); } } let repo = Repo::open(config.repo.clone()).await?; if config.server.read_only { tracing::warn!("Launching in READ ONLY mode"); } match config.store.clone() { config::Store::Filesystem(config::Filesystem { path }) => { let arc_repo = repo.to_arc(); let store = FileStore::build(path, arc_repo.clone()).await?; if arc_repo.get("migrate-0.4").await?.is_none() { if let Some(path) = config.old_repo_path() { if let Some(old_repo) = repo_04::open(path)? { repo::migrate_04( tmp_dir.clone(), old_repo, arc_repo.clone(), store.clone(), config.clone(), ) .await?; arc_repo .set("migrate-0.4", Arc::from(b"migrated".to_vec())) .await?; } } } match repo { Repo::Sled(sled_repo) => { launch_file_store( tmp_dir.clone(), arc_repo, store, client, config, move |sc| sled_extra_config(sc, sled_repo.clone()), ) .await?; } Repo::Postgres(_) => { launch_file_store(tmp_dir.clone(), arc_repo, store, client, config, |_| {}) .await?; } } } config::Store::ObjectStorage(config::ObjectStorage { endpoint, bucket_name, use_path_style, region, access_key, secret_key, session_token, signature_duration, client_timeout, public_endpoint, }) => { let arc_repo = repo.to_arc(); let store = ObjectStore::build( endpoint, bucket_name, if use_path_style { UrlStyle::Path } else { UrlStyle::VirtualHost }, region, access_key, secret_key, session_token, signature_duration, client_timeout, public_endpoint, arc_repo.clone(), ) .await? .build(client.clone()); if arc_repo.get("migrate-0.4").await?.is_none() { if let Some(path) = config.old_repo_path() { if let Some(old_repo) = repo_04::open(path)? { repo::migrate_04( tmp_dir.clone(), old_repo, arc_repo.clone(), store.clone(), config.clone(), ) .await?; arc_repo .set("migrate-0.4", Arc::from(b"migrated".to_vec())) .await?; } } } match repo { Repo::Sled(sled_repo) => { launch_object_store( tmp_dir.clone(), arc_repo, store, client, config, move |sc| sled_extra_config(sc, sled_repo.clone()), ) .await?; } Repo::Postgres(_) => { launch_object_store( tmp_dir.clone(), arc_repo, store, client, config, |_| {}, ) .await?; } } } } tmp_dir.cleanup().await?; Ok(()) } } #[cfg(test)] mod tests { #[test] fn source() { let query = super::ProcessQuery { source: super::ProcessSource::Source { src: super::Serde::new(super::Alias::from_existing("example.png")), }, operations: vec![("resize".into(), "200".into())], }; let encoded = serde_urlencoded::to_string(&query).expect("Encoded"); let new_query: super::ProcessQuery = serde_urlencoded::from_str(&encoded).expect("Decoded"); // Don't compare entire query - "src" gets deserialized twice assert_eq!(new_query.source, query.source); assert!(new_query .operations .contains(&("resize".into(), "200".into()))); } #[test] fn alias() { let query = super::ProcessQuery { source: super::ProcessSource::Alias { alias: super::Serde::new(super::Alias::from_existing("example.png")), }, operations: vec![("resize".into(), "200".into())], }; let encoded = serde_urlencoded::to_string(&query).expect("Encoded"); let new_query: super::ProcessQuery = serde_urlencoded::from_str(&encoded).expect("Decoded"); // Don't compare entire query - "alias" gets deserialized twice assert_eq!(new_query.source, query.source); assert!(new_query .operations .contains(&("resize".into(), "200".into()))); } #[test] fn url() { let query = super::ProcessQuery { source: super::ProcessSource::Proxy { proxy: "http://example.com/image.png".parse().expect("valid url"), }, operations: vec![("resize".into(), "200".into())], }; let encoded = serde_urlencoded::to_string(&query).expect("Encoded"); let new_query: super::ProcessQuery = serde_urlencoded::from_str(&encoded).expect("Decoded"); // Don't compare entire query - "proxy" gets deserialized twice assert_eq!(new_query.source, query.source); assert!(new_query .operations .contains(&("resize".into(), "200".into()))); } }