use crate::{ config, details::Details, store::{file_store::FileId, Identifier, StoreError}, }; use base64::{prelude::BASE64_STANDARD, Engine}; use futures_util::Stream; use std::{fmt::Debug, path::PathBuf}; use tracing::Instrument; use uuid::Uuid; mod old; pub(crate) mod sled; #[derive(Clone, Debug)] pub(crate) enum Repo { Sled(self::sled::SledRepo), } #[derive(Clone, Debug, PartialEq, Eq, PartialOrd, Ord, Hash)] enum MaybeUuid { Uuid(Uuid), Name(String), } #[derive(Clone, Debug, PartialEq, Eq, PartialOrd, Ord, Hash)] pub(crate) struct Alias { id: MaybeUuid, extension: Option, } #[derive(Debug, PartialEq, Eq, PartialOrd, Ord, Hash)] pub(crate) struct DeleteToken { id: MaybeUuid, } pub(crate) struct HashAlreadyExists; pub(crate) struct AliasAlreadyExists; pub(crate) struct AlreadyExists(pub(super) DeleteToken); #[derive(Clone, Copy, Debug, PartialEq, Eq, PartialOrd, Ord, Hash)] pub(crate) struct UploadId { id: Uuid, } pub(crate) enum UploadResult { Success { alias: Alias, token: DeleteToken }, Failure { message: String }, } #[derive(Debug, thiserror::Error)] pub(crate) enum RepoError { #[error("Error in sled")] SledError(#[from] crate::repo::sled::SledError), #[error("Upload was already claimed")] AlreadyClaimed, #[error("Panic in blocking operation")] Canceled, #[error("Required field is missing {0}")] Missing(&'static str), } #[async_trait::async_trait(?Send)] pub(crate) trait FullRepo: UploadRepo + SettingsRepo + IdentifierRepo + AliasRepo + QueueRepo + HashRepo + MigrationRepo + Send + Sync + Clone + Debug { async fn health_check(&self) -> Result<(), RepoError>; #[tracing::instrument(skip(self))] async fn identifier_from_alias( &self, alias: &Alias, ) -> Result, StoreError> { let Some(hash) = self.hash(alias).await? else { return Ok(None); }; self.identifier(hash).await } #[tracing::instrument(skip(self))] async fn aliases_from_alias(&self, alias: &Alias) -> Result, RepoError> { let Some(hash) = self.hash(alias).await? else { return Ok(vec![]); }; self.aliases(hash).await } #[tracing::instrument(skip(self))] async fn still_identifier_from_alias( &self, alias: &Alias, ) -> Result, StoreError> { let Some(hash) = self.hash(alias).await? else { return Ok(None); }; let Some(identifier) = self.identifier::(hash.clone()).await? else { return Ok(None); }; match self.details(&identifier).await? { Some(details) if details.is_motion() => self.motion_identifier::(hash).await, Some(_) => Ok(Some(identifier)), None => Ok(None), } } } #[async_trait::async_trait(?Send)] impl FullRepo for actix_web::web::Data where T: FullRepo, { async fn health_check(&self) -> Result<(), RepoError> { T::health_check(self).await } } pub(crate) trait BaseRepo { type Bytes: AsRef<[u8]> + From> + Clone; } impl BaseRepo for actix_web::web::Data where T: BaseRepo, { type Bytes = T::Bytes; } #[async_trait::async_trait(?Send)] pub(crate) trait UploadRepo: BaseRepo { async fn create(&self, upload_id: UploadId) -> Result<(), RepoError>; async fn wait(&self, upload_id: UploadId) -> Result; async fn claim(&self, upload_id: UploadId) -> Result<(), RepoError>; async fn complete(&self, upload_id: UploadId, result: UploadResult) -> Result<(), RepoError>; } #[async_trait::async_trait(?Send)] impl UploadRepo for actix_web::web::Data where T: UploadRepo, { async fn create(&self, upload_id: UploadId) -> Result<(), RepoError> { T::create(self, upload_id).await } async fn wait(&self, upload_id: UploadId) -> Result { T::wait(self, upload_id).await } async fn claim(&self, upload_id: UploadId) -> Result<(), RepoError> { T::claim(self, upload_id).await } async fn complete(&self, upload_id: UploadId, result: UploadResult) -> Result<(), RepoError> { T::complete(self, upload_id, result).await } } #[async_trait::async_trait(?Send)] pub(crate) trait QueueRepo: BaseRepo { async fn requeue_in_progress(&self, worker_prefix: Vec) -> Result<(), RepoError>; async fn push(&self, queue: &'static str, job: Self::Bytes) -> Result<(), RepoError>; async fn pop(&self, queue: &'static str, worker_id: Vec) -> Result; } #[async_trait::async_trait(?Send)] impl QueueRepo for actix_web::web::Data where T: QueueRepo, { async fn requeue_in_progress(&self, worker_prefix: Vec) -> Result<(), RepoError> { T::requeue_in_progress(self, worker_prefix).await } async fn push(&self, queue: &'static str, job: Self::Bytes) -> Result<(), RepoError> { T::push(self, queue, job).await } async fn pop(&self, queue: &'static str, worker_id: Vec) -> Result { T::pop(self, queue, worker_id).await } } #[async_trait::async_trait(?Send)] pub(crate) trait SettingsRepo: BaseRepo { async fn set(&self, key: &'static str, value: Self::Bytes) -> Result<(), RepoError>; async fn get(&self, key: &'static str) -> Result, RepoError>; async fn remove(&self, key: &'static str) -> Result<(), RepoError>; } #[async_trait::async_trait(?Send)] impl SettingsRepo for actix_web::web::Data where T: SettingsRepo, { async fn set(&self, key: &'static str, value: Self::Bytes) -> Result<(), RepoError> { T::set(self, key, value).await } async fn get(&self, key: &'static str) -> Result, RepoError> { T::get(self, key).await } async fn remove(&self, key: &'static str) -> Result<(), RepoError> { T::remove(self, key).await } } #[async_trait::async_trait(?Send)] pub(crate) trait IdentifierRepo: BaseRepo { async fn relate_details( &self, identifier: &I, details: &Details, ) -> Result<(), StoreError>; async fn details(&self, identifier: &I) -> Result, StoreError>; async fn cleanup(&self, identifier: &I) -> Result<(), StoreError>; } #[async_trait::async_trait(?Send)] impl IdentifierRepo for actix_web::web::Data where T: IdentifierRepo, { async fn relate_details( &self, identifier: &I, details: &Details, ) -> Result<(), StoreError> { T::relate_details(self, identifier, details).await } async fn details(&self, identifier: &I) -> Result, StoreError> { T::details(self, identifier).await } async fn cleanup(&self, identifier: &I) -> Result<(), StoreError> { T::cleanup(self, identifier).await } } #[async_trait::async_trait(?Send)] pub(crate) trait MigrationRepo: BaseRepo { async fn is_continuing_migration(&self) -> Result; async fn mark_migrated( &self, old_identifier: &I1, new_identifier: &I2, ) -> Result<(), StoreError>; async fn is_migrated(&self, identifier: &I) -> Result; async fn clear(&self) -> Result<(), RepoError>; } #[async_trait::async_trait(?Send)] impl MigrationRepo for actix_web::web::Data where T: MigrationRepo, { async fn is_continuing_migration(&self) -> Result { T::is_continuing_migration(self).await } async fn mark_migrated( &self, old_identifier: &I1, new_identifier: &I2, ) -> Result<(), StoreError> { T::mark_migrated(self, old_identifier, new_identifier).await } async fn is_migrated(&self, identifier: &I) -> Result { T::is_migrated(self, identifier).await } async fn clear(&self) -> Result<(), RepoError> { T::clear(self).await } } #[async_trait::async_trait(?Send)] pub(crate) trait HashRepo: BaseRepo { type Stream: Stream>; async fn size(&self) -> Result; async fn hashes(&self) -> Self::Stream; async fn create(&self, hash: Self::Bytes) -> Result, RepoError>; async fn relate_alias(&self, hash: Self::Bytes, alias: &Alias) -> Result<(), RepoError>; async fn remove_alias(&self, hash: Self::Bytes, alias: &Alias) -> Result<(), RepoError>; async fn aliases(&self, hash: Self::Bytes) -> Result, RepoError>; async fn relate_identifier( &self, hash: Self::Bytes, identifier: &I, ) -> Result<(), StoreError>; async fn identifier( &self, hash: Self::Bytes, ) -> Result, StoreError>; async fn relate_variant_identifier( &self, hash: Self::Bytes, variant: String, identifier: &I, ) -> Result<(), StoreError>; async fn variant_identifier( &self, hash: Self::Bytes, variant: String, ) -> Result, StoreError>; async fn variants( &self, hash: Self::Bytes, ) -> Result, StoreError>; async fn remove_variant(&self, hash: Self::Bytes, variant: String) -> Result<(), RepoError>; async fn relate_motion_identifier( &self, hash: Self::Bytes, identifier: &I, ) -> Result<(), StoreError>; async fn motion_identifier( &self, hash: Self::Bytes, ) -> Result, StoreError>; async fn cleanup(&self, hash: Self::Bytes) -> Result<(), RepoError>; } #[async_trait::async_trait(?Send)] impl HashRepo for actix_web::web::Data where T: HashRepo, { type Stream = T::Stream; async fn size(&self) -> Result { T::size(self).await } async fn hashes(&self) -> Self::Stream { T::hashes(self).await } async fn create(&self, hash: Self::Bytes) -> Result, RepoError> { T::create(self, hash).await } async fn relate_alias(&self, hash: Self::Bytes, alias: &Alias) -> Result<(), RepoError> { T::relate_alias(self, hash, alias).await } async fn remove_alias(&self, hash: Self::Bytes, alias: &Alias) -> Result<(), RepoError> { T::remove_alias(self, hash, alias).await } async fn aliases(&self, hash: Self::Bytes) -> Result, RepoError> { T::aliases(self, hash).await } async fn relate_identifier( &self, hash: Self::Bytes, identifier: &I, ) -> Result<(), StoreError> { T::relate_identifier(self, hash, identifier).await } async fn identifier( &self, hash: Self::Bytes, ) -> Result, StoreError> { T::identifier(self, hash).await } async fn relate_variant_identifier( &self, hash: Self::Bytes, variant: String, identifier: &I, ) -> Result<(), StoreError> { T::relate_variant_identifier(self, hash, variant, identifier).await } async fn variant_identifier( &self, hash: Self::Bytes, variant: String, ) -> Result, StoreError> { T::variant_identifier(self, hash, variant).await } async fn variants( &self, hash: Self::Bytes, ) -> Result, StoreError> { T::variants(self, hash).await } async fn remove_variant(&self, hash: Self::Bytes, variant: String) -> Result<(), RepoError> { T::remove_variant(self, hash, variant).await } async fn relate_motion_identifier( &self, hash: Self::Bytes, identifier: &I, ) -> Result<(), StoreError> { T::relate_motion_identifier(self, hash, identifier).await } async fn motion_identifier( &self, hash: Self::Bytes, ) -> Result, StoreError> { T::motion_identifier(self, hash).await } async fn cleanup(&self, hash: Self::Bytes) -> Result<(), RepoError> { T::cleanup(self, hash).await } } #[async_trait::async_trait(?Send)] pub(crate) trait AliasRepo: BaseRepo { async fn create(&self, alias: &Alias) -> Result, RepoError>; async fn relate_delete_token( &self, alias: &Alias, delete_token: &DeleteToken, ) -> Result, RepoError>; async fn delete_token(&self, alias: &Alias) -> Result, RepoError>; async fn relate_hash(&self, alias: &Alias, hash: Self::Bytes) -> Result<(), RepoError>; async fn hash(&self, alias: &Alias) -> Result, RepoError>; async fn cleanup(&self, alias: &Alias) -> Result<(), RepoError>; } #[async_trait::async_trait(?Send)] impl AliasRepo for actix_web::web::Data where T: AliasRepo, { async fn create(&self, alias: &Alias) -> Result, RepoError> { T::create(self, alias).await } async fn relate_delete_token( &self, alias: &Alias, delete_token: &DeleteToken, ) -> Result, RepoError> { T::relate_delete_token(self, alias, delete_token).await } async fn delete_token(&self, alias: &Alias) -> Result, RepoError> { T::delete_token(self, alias).await } async fn relate_hash(&self, alias: &Alias, hash: Self::Bytes) -> Result<(), RepoError> { T::relate_hash(self, alias, hash).await } async fn hash(&self, alias: &Alias) -> Result, RepoError> { T::hash(self, alias).await } async fn cleanup(&self, alias: &Alias) -> Result<(), RepoError> { T::cleanup(self, alias).await } } impl Repo { #[tracing::instrument] pub(crate) fn open(config: config::Repo) -> color_eyre::Result { match config { config::Repo::Sled(config::Sled { path, cache_capacity, export_path, }) => { let repo = self::sled::SledRepo::build(path, cache_capacity, export_path)?; Ok(Self::Sled(repo)) } } } pub(crate) async fn migrate_from_db(&self, path: PathBuf) -> color_eyre::Result<()> { if self.has_migrated().await? { return Ok(()); } if let Some(old) = self::old::Old::open(path)? { let span = tracing::warn_span!("Migrating Database from 0.3 layout to 0.4 layout"); match self { Self::Sled(repo) => { async { for hash in old.hashes() { if let Err(e) = migrate_hash(repo, &old, hash).await { tracing::error!("Failed to migrate hash: {}", format!("{e:?}")); } } if let Ok(Some(value)) = old.setting(STORE_MIGRATION_PROGRESS.as_bytes()) { tracing::warn!("Setting STORE_MIGRATION_PROGRESS"); let _ = repo.set(STORE_MIGRATION_PROGRESS, value).await; } if let Ok(Some(value)) = old.setting(GENERATOR_KEY.as_bytes()) { tracing::warn!("Setting GENERATOR_KEY"); let _ = repo.set(GENERATOR_KEY, value).await; } } .instrument(span) .await; } } } self.mark_migrated().await?; Ok(()) } pub(crate) async fn migrate_identifiers(&self) -> color_eyre::Result<()> { if self.has_migrated_identifiers().await? { return Ok(()); } let span = tracing::warn_span!("Migrating File Identifiers from 0.3 format to 0.4 format"); match self { Self::Sled(repo) => { async { use futures_util::StreamExt; let mut hashes = repo.hashes().await; while let Some(res) = hashes.next().await { let hash = res?; if let Err(e) = migrate_identifiers_for_hash(repo, hash).await { tracing::error!( "Failed to migrate identifiers for hash: {}", format!("{e:?}") ); } } Ok(()) as color_eyre::Result<()> } .instrument(span) .await?; } } self.mark_migrated_identifiers().await?; Ok(()) } async fn has_migrated(&self) -> color_eyre::Result { match self { Self::Sled(repo) => Ok(repo.get(REPO_MIGRATION_O1).await?.is_some()), } } async fn has_migrated_identifiers(&self) -> color_eyre::Result { match self { Self::Sled(repo) => Ok(repo.get(REPO_MIGRATION_02).await?.is_some()), } } async fn mark_migrated(&self) -> color_eyre::Result<()> { match self { Self::Sled(repo) => { repo.set(REPO_MIGRATION_O1, b"1".to_vec().into()).await?; } } Ok(()) } async fn mark_migrated_identifiers(&self) -> color_eyre::Result<()> { match self { Self::Sled(repo) => { repo.set(REPO_MIGRATION_02, b"1".to_vec().into()).await?; } } Ok(()) } } const REPO_MIGRATION_O1: &str = "repo-migration-01"; const REPO_MIGRATION_02: &str = "repo-migration-02"; const STORE_MIGRATION_PROGRESS: &str = "store-migration-progress"; const GENERATOR_KEY: &str = "last-path"; #[tracing::instrument(skip(repo, hash), fields(hash = hex::encode(&hash)))] async fn migrate_identifiers_for_hash(repo: &T, hash: ::sled::IVec) -> color_eyre::Result<()> where T: FullRepo, { let hash: T::Bytes = hash.to_vec().into(); if let Some(motion_identifier) = repo.motion_identifier::(hash.clone()).await? { if let Some(new_motion_identifier) = motion_identifier.normalize_for_migration() { migrate_identifier_details(repo, &motion_identifier, &new_motion_identifier).await?; repo.relate_motion_identifier(hash.clone(), &new_motion_identifier) .await?; } } for (variant_path, variant_identifier) in repo.variants::(hash.clone()).await? { if let Some(new_variant_identifier) = variant_identifier.normalize_for_migration() { migrate_identifier_details(repo, &variant_identifier, &new_variant_identifier).await?; repo.relate_variant_identifier(hash.clone(), variant_path, &new_variant_identifier) .await?; } } let Some(main_identifier) = repo.identifier::(hash.clone()).await? else { tracing::warn!("Missing identifier for hash {}, queueing cleanup", hex::encode(&hash)); crate::queue::cleanup_hash(repo, hash.clone()).await?; return Err(RepoError::Missing("hash -> identifier").into()); }; if let Some(new_main_identifier) = main_identifier.normalize_for_migration() { migrate_identifier_details(repo, &main_identifier, &new_main_identifier).await?; repo.relate_identifier(hash, &new_main_identifier).await?; } Ok(()) } #[tracing::instrument(level = "debug", skip(repo))] async fn migrate_identifier_details( repo: &T, old: &FileId, new: &FileId, ) -> color_eyre::Result<()> where T: FullRepo, { if old == new { tracing::warn!("Old FileId and new FileId are identical"); return Ok(()); } if let Some(details) = repo.details(old).await? { repo.relate_details(new, &details).await?; IdentifierRepo::cleanup(repo, old).await?; } Ok(()) } #[tracing::instrument(skip(repo, old, hash), fields(hash = hex::encode(&hash)))] async fn migrate_hash(repo: &T, old: &old::Old, hash: ::sled::IVec) -> color_eyre::Result<()> where T: IdentifierRepo + HashRepo + AliasRepo + SettingsRepo + Debug, { let new_hash: T::Bytes = hash.to_vec().into(); let main_ident = old.main_identifier(&hash)?.to_vec(); if HashRepo::create(repo, new_hash.clone()).await?.is_err() { tracing::warn!("Duplicate hash detected"); return Ok(()); } repo.relate_identifier(new_hash.clone(), &main_ident) .await?; for alias in old.aliases(&hash) { if let Ok(Ok(())) = AliasRepo::create(repo, &alias).await { let _ = repo.relate_alias(new_hash.clone(), &alias).await; let _ = repo.relate_hash(&alias, new_hash.clone()).await; if let Ok(Some(delete_token)) = old.delete_token(&alias) { let _ = repo.relate_delete_token(&alias, &delete_token).await; } } } if let Ok(Some(identifier)) = old.motion_identifier(&hash) { let _ = repo .relate_motion_identifier(new_hash.clone(), &identifier.to_vec()) .await; } for (variant_path, identifier) in old.variants(&hash)? { let variant = variant_path.to_string_lossy().to_string(); let _ = repo .relate_variant_identifier(new_hash.clone(), variant, &identifier.to_vec()) .await; } for (identifier, details) in old.details(&hash)? { let _ = repo.relate_details(&identifier.to_vec(), &details).await; } Ok(()) } impl MaybeUuid { fn from_str(s: &str) -> Self { if let Ok(uuid) = Uuid::parse_str(s) { MaybeUuid::Uuid(uuid) } else { MaybeUuid::Name(s.into()) } } fn as_bytes(&self) -> &[u8] { match self { Self::Uuid(uuid) => &uuid.as_bytes()[..], Self::Name(name) => name.as_bytes(), } } } fn split_at_dot(s: &str) -> Option<(&str, &str)> { let index = s.find('.')?; Some(s.split_at(index)) } impl Alias { pub(crate) fn generate(extension: String) -> Self { Alias { id: MaybeUuid::Uuid(Uuid::new_v4()), extension: Some(extension), } } pub(crate) fn from_existing(alias: &str) -> Self { if let Some((start, end)) = split_at_dot(alias) { Alias { id: MaybeUuid::from_str(start), extension: Some(end.into()), } } else { Alias { id: MaybeUuid::from_str(alias), extension: None, } } } pub(crate) fn extension(&self) -> Option<&str> { self.extension.as_deref() } pub(crate) fn to_bytes(&self) -> Vec { let mut v = self.id.as_bytes().to_vec(); if let Some(ext) = self.extension() { v.extend_from_slice(ext.as_bytes()); } v } pub(crate) fn from_slice(bytes: &[u8]) -> Option { if let Ok(s) = std::str::from_utf8(bytes) { Some(Self::from_existing(s)) } else if bytes.len() >= 16 { let id = Uuid::from_slice(&bytes[0..16]).expect("Already checked length"); let extension = if bytes.len() > 16 { Some(String::from_utf8_lossy(&bytes[16..]).to_string()) } else { None }; Some(Self { id: MaybeUuid::Uuid(id), extension, }) } else { None } } } impl DeleteToken { pub(crate) fn from_existing(existing: &str) -> Self { if let Ok(uuid) = Uuid::parse_str(existing) { DeleteToken { id: MaybeUuid::Uuid(uuid), } } else { DeleteToken { id: MaybeUuid::Name(existing.into()), } } } pub(crate) fn generate() -> Self { Self { id: MaybeUuid::Uuid(Uuid::new_v4()), } } fn to_bytes(&self) -> Vec { self.id.as_bytes().to_vec() } fn from_slice(bytes: &[u8]) -> Option { if let Ok(s) = std::str::from_utf8(bytes) { Some(DeleteToken::from_existing(s)) } else if bytes.len() == 16 { Some(DeleteToken { id: MaybeUuid::Uuid(Uuid::from_slice(bytes).ok()?), }) } else { None } } } impl UploadId { pub(crate) fn generate() -> Self { Self { id: Uuid::new_v4() } } pub(crate) fn as_bytes(&self) -> &[u8] { &self.id.as_bytes()[..] } } impl std::str::FromStr for UploadId { type Err = ::Err; fn from_str(s: &str) -> Result { Ok(UploadId { id: s.parse()? }) } } impl std::fmt::Display for UploadId { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { std::fmt::Display::fmt(&self.id, f) } } impl std::fmt::Display for MaybeUuid { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { match self { Self::Uuid(id) => write!(f, "{id}"), Self::Name(name) => write!(f, "{name}"), } } } impl std::str::FromStr for DeleteToken { type Err = std::convert::Infallible; fn from_str(s: &str) -> Result { Ok(DeleteToken::from_existing(s)) } } impl std::fmt::Display for DeleteToken { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { write!(f, "{}", self.id) } } impl std::str::FromStr for Alias { type Err = std::convert::Infallible; fn from_str(s: &str) -> Result { Ok(Alias::from_existing(s)) } } impl std::fmt::Display for Alias { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { if let Some(ext) = self.extension() { write!(f, "{}{ext}", self.id) } else { write!(f, "{}", self.id) } } } impl Identifier for Vec { fn from_bytes(bytes: Vec) -> Result where Self: Sized, { Ok(bytes) } fn to_bytes(&self) -> Result, StoreError> { Ok(self.clone()) } fn string_repr(&self) -> String { BASE64_STANDARD.encode(self.as_slice()) } } #[cfg(test)] mod tests { use super::{Alias, DeleteToken, MaybeUuid, Uuid}; #[test] fn string_delete_token() { let delete_token = DeleteToken::from_existing("blah"); assert_eq!( delete_token, DeleteToken { id: MaybeUuid::Name(String::from("blah")) } ) } #[test] fn uuid_string_delete_token() { let uuid = Uuid::new_v4(); let delete_token = DeleteToken::from_existing(&uuid.to_string()); assert_eq!( delete_token, DeleteToken { id: MaybeUuid::Uuid(uuid), } ) } #[test] fn bytes_delete_token() { let delete_token = DeleteToken::from_slice(b"blah").unwrap(); assert_eq!( delete_token, DeleteToken { id: MaybeUuid::Name(String::from("blah")) } ) } #[test] fn uuid_bytes_delete_token() { let uuid = Uuid::new_v4(); let delete_token = DeleteToken::from_slice(&uuid.as_bytes()[..]).unwrap(); assert_eq!( delete_token, DeleteToken { id: MaybeUuid::Uuid(uuid), } ) } #[test] fn uuid_bytes_string_delete_token() { let uuid = Uuid::new_v4(); let delete_token = DeleteToken::from_slice(uuid.to_string().as_bytes()).unwrap(); assert_eq!( delete_token, DeleteToken { id: MaybeUuid::Uuid(uuid), } ) } #[test] fn string_alias() { let alias = Alias::from_existing("blah"); assert_eq!( alias, Alias { id: MaybeUuid::Name(String::from("blah")), extension: None } ); } #[test] fn string_alias_ext() { let alias = Alias::from_existing("blah.mp4"); assert_eq!( alias, Alias { id: MaybeUuid::Name(String::from("blah")), extension: Some(String::from(".mp4")), } ); } #[test] fn uuid_string_alias() { let uuid = Uuid::new_v4(); let alias = Alias::from_existing(&uuid.to_string()); assert_eq!( alias, Alias { id: MaybeUuid::Uuid(uuid), extension: None, } ) } #[test] fn uuid_string_alias_ext() { let uuid = Uuid::new_v4(); let alias_str = format!("{uuid}.mp4"); let alias = Alias::from_existing(&alias_str); assert_eq!( alias, Alias { id: MaybeUuid::Uuid(uuid), extension: Some(String::from(".mp4")), } ) } #[test] fn bytes_alias() { let alias = Alias::from_slice(b"blah").unwrap(); assert_eq!( alias, Alias { id: MaybeUuid::Name(String::from("blah")), extension: None } ); } #[test] fn bytes_alias_ext() { let alias = Alias::from_slice(b"blah.mp4").unwrap(); assert_eq!( alias, Alias { id: MaybeUuid::Name(String::from("blah")), extension: Some(String::from(".mp4")), } ); } #[test] fn uuid_bytes_alias() { let uuid = Uuid::new_v4(); let alias = Alias::from_slice(&uuid.as_bytes()[..]).unwrap(); assert_eq!( alias, Alias { id: MaybeUuid::Uuid(uuid), extension: None, } ) } #[test] fn uuid_bytes_string_alias() { let uuid = Uuid::new_v4(); let alias = Alias::from_slice(uuid.to_string().as_bytes()).unwrap(); assert_eq!( alias, Alias { id: MaybeUuid::Uuid(uuid), extension: None, } ) } #[test] fn uuid_bytes_alias_ext() { let uuid = Uuid::new_v4(); let mut alias_bytes = uuid.as_bytes().to_vec(); alias_bytes.extend_from_slice(b".mp4"); let alias = Alias::from_slice(&alias_bytes).unwrap(); assert_eq!( alias, Alias { id: MaybeUuid::Uuid(uuid), extension: Some(String::from(".mp4")), } ) } #[test] fn uuid_bytes_string_alias_ext() { let uuid = Uuid::new_v4(); let alias_str = format!("{uuid}.mp4"); let alias = Alias::from_slice(alias_str.as_bytes()).unwrap(); assert_eq!( alias, Alias { id: MaybeUuid::Uuid(uuid), extension: Some(String::from(".mp4")), } ) } }