Compare commits
No commits in common. "28c2884a2e08d33d7dd50c23a4dd2b5e7bc4d912" and "54fe339ce5fc7e8fbb8935c927a7ec0e54b17750" have entirely different histories.
28c2884a2e
...
54fe339ce5
16 changed files with 70 additions and 961 deletions
69
Cargo.lock
generated
69
Cargo.lock
generated
|
@ -17,21 +17,6 @@ dependencies = [
|
||||||
"memchr",
|
"memchr",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "alloc-no-stdlib"
|
|
||||||
version = "2.0.4"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "cc7bb162ec39d46ab1ca8c77bf72e890535becd1751bb45f64c597edb4c8c6b3"
|
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "alloc-stdlib"
|
|
||||||
version = "0.2.2"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "94fb8275041c72129eb51b7d0322c29b8387a0386127718b096429201a5d6ece"
|
|
||||||
dependencies = [
|
|
||||||
"alloc-no-stdlib",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "async-trait"
|
name = "async-trait"
|
||||||
version = "0.1.64"
|
version = "0.1.64"
|
||||||
|
@ -70,27 +55,6 @@ dependencies = [
|
||||||
"generic-array",
|
"generic-array",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "brotli"
|
|
||||||
version = "3.3.4"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "a1a0b1dbcc8ae29329621f8d4f0d835787c1c38bb1401979b49d13b0b305ff68"
|
|
||||||
dependencies = [
|
|
||||||
"alloc-no-stdlib",
|
|
||||||
"alloc-stdlib",
|
|
||||||
"brotli-decompressor",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "brotli-decompressor"
|
|
||||||
version = "2.3.4"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "4b6561fd3f895a11e8f72af2cb7d22e08366bebc2b6b57f7744c4bda27034744"
|
|
||||||
dependencies = [
|
|
||||||
"alloc-no-stdlib",
|
|
||||||
"alloc-stdlib",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "byteorder"
|
name = "byteorder"
|
||||||
version = "1.4.3"
|
version = "1.4.3"
|
||||||
|
@ -139,15 +103,6 @@ version = "1.0.0"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "baf1de4339761588bc0619e3cbc0120ee582ebb74b53b4efbf79117bd2da40fd"
|
checksum = "baf1de4339761588bc0619e3cbc0120ee582ebb74b53b4efbf79117bd2da40fd"
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "compressible"
|
|
||||||
version = "0.2.0"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "fe7853faa8a7c37cddc40823bf5463d368a5207ebb4e7d4d83846da656f493d3"
|
|
||||||
dependencies = [
|
|
||||||
"mime",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "console"
|
name = "console"
|
||||||
version = "0.15.5"
|
version = "0.15.5"
|
||||||
|
@ -617,16 +572,6 @@ version = "0.3.16"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "2a60c7ce501c71e03a9c9c0d35b861413ae925bd979cc7a4e30d060069aaac8d"
|
checksum = "2a60c7ce501c71e03a9c9c0d35b861413ae925bd979cc7a4e30d060069aaac8d"
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "mime_guess"
|
|
||||||
version = "2.0.4"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "4192263c238a5f0d0c6bfd21f336a313a4ce1c450542449ca191bb657b4642ef"
|
|
||||||
dependencies = [
|
|
||||||
"mime",
|
|
||||||
"unicase",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "miniz_oxide"
|
name = "miniz_oxide"
|
||||||
version = "0.6.2"
|
version = "0.6.2"
|
||||||
|
@ -1082,17 +1027,12 @@ dependencies = [
|
||||||
name = "talon"
|
name = "talon"
|
||||||
version = "0.1.0"
|
version = "0.1.0"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"brotli",
|
|
||||||
"compressible",
|
|
||||||
"flate2",
|
"flate2",
|
||||||
"hex",
|
"hex",
|
||||||
"hex-literal",
|
"hex-literal",
|
||||||
"insta",
|
"insta",
|
||||||
"log",
|
|
||||||
"mime_guess",
|
|
||||||
"path_macro",
|
"path_macro",
|
||||||
"poem",
|
"poem",
|
||||||
"regex",
|
|
||||||
"rmp-serde",
|
"rmp-serde",
|
||||||
"rstest",
|
"rstest",
|
||||||
"serde",
|
"serde",
|
||||||
|
@ -1330,15 +1270,6 @@ dependencies = [
|
||||||
"version_check",
|
"version_check",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "unicase"
|
|
||||||
version = "2.6.0"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "50f37be617794602aabbeee0be4f259dc1778fabe05e2d67ee8f79326d5cb4f6"
|
|
||||||
dependencies = [
|
|
||||||
"version_check",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "unicode-ident"
|
name = "unicode-ident"
|
||||||
version = "1.0.6"
|
version = "1.0.6"
|
||||||
|
|
|
@ -12,7 +12,7 @@ sled = "0.34.7"
|
||||||
serde = "1.0.152"
|
serde = "1.0.152"
|
||||||
serde_json = "1.0.93"
|
serde_json = "1.0.93"
|
||||||
rmp-serde = "1.1.1"
|
rmp-serde = "1.1.1"
|
||||||
toml = { version = "0.7.2", default-features = false, features = ["parse"] }
|
toml = "0.7.2"
|
||||||
thiserror = "1.0.38"
|
thiserror = "1.0.38"
|
||||||
time = { version = "0.3.15", features = [
|
time = { version = "0.3.15", features = [
|
||||||
"macros",
|
"macros",
|
||||||
|
@ -30,11 +30,6 @@ zip = { version = "0.6.4", default-features = false, features = [
|
||||||
] }
|
] }
|
||||||
tar = "0.4.38"
|
tar = "0.4.38"
|
||||||
flate2 = "1.0.25"
|
flate2 = "1.0.25"
|
||||||
brotli = "3.3.4"
|
|
||||||
mime_guess = { version = "2.0.4", default-features = false }
|
|
||||||
compressible = "0.2.0"
|
|
||||||
regex = "1.7.1"
|
|
||||||
log = "0.4.17"
|
|
||||||
|
|
||||||
[dev-dependencies]
|
[dev-dependencies]
|
||||||
rstest = "0.16.0"
|
rstest = "0.16.0"
|
||||||
|
|
177
src/config.rs
177
src/config.rs
|
@ -1,177 +0,0 @@
|
||||||
use std::{collections::BTreeMap, ops::Deref, path::Path, sync::Arc};
|
|
||||||
|
|
||||||
use regex::Regex;
|
|
||||||
use serde::{Deserialize, Serialize};
|
|
||||||
|
|
||||||
#[derive(Clone, Default)]
|
|
||||||
pub struct Config {
|
|
||||||
i: Arc<ConfigInner>,
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Debug, Default, Clone, Serialize, Deserialize)]
|
|
||||||
#[serde(default)]
|
|
||||||
pub struct ConfigInner {
|
|
||||||
pub server: ServerCfg,
|
|
||||||
pub compression: CompressionCfg,
|
|
||||||
pub keys: BTreeMap<String, KeyCfg>,
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(thiserror::Error, Debug)]
|
|
||||||
pub enum ConfigError {
|
|
||||||
#[error("io error: {0}")]
|
|
||||||
Io(#[from] std::io::Error),
|
|
||||||
#[error("parsing error: {0}")]
|
|
||||||
Parse(#[from] toml::de::Error),
|
|
||||||
}
|
|
||||||
|
|
||||||
type Result<T> = std::result::Result<T, ConfigError>;
|
|
||||||
|
|
||||||
impl Deref for Config {
|
|
||||||
type Target = ConfigInner;
|
|
||||||
|
|
||||||
fn deref(&self) -> &Self::Target {
|
|
||||||
&self.i
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl Serialize for Config {
|
|
||||||
fn serialize<S>(&self, serializer: S) -> std::result::Result<S::Ok, S::Error>
|
|
||||||
where
|
|
||||||
S: serde::Serializer,
|
|
||||||
{
|
|
||||||
ConfigInner::serialize(self, serializer)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl<'de> Deserialize<'de> for Config {
|
|
||||||
fn deserialize<D>(deserializer: D) -> std::result::Result<Self, D::Error>
|
|
||||||
where
|
|
||||||
D: serde::Deserializer<'de>,
|
|
||||||
{
|
|
||||||
ConfigInner::deserialize(deserializer).map(|c| Self { i: c.into() })
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl Config {
|
|
||||||
pub fn new(cfg: ConfigInner) -> Self {
|
|
||||||
Self { i: cfg.into() }
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn from_file<P: AsRef<Path>>(path: P) -> Result<Self> {
|
|
||||||
let cfg_str = std::fs::read_to_string(path)?;
|
|
||||||
Ok(toml::from_str::<Config>(&cfg_str)?)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
|
||||||
#[serde(default)]
|
|
||||||
pub struct ServerCfg {
|
|
||||||
pub address: String,
|
|
||||||
pub port: u32,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl Default for ServerCfg {
|
|
||||||
fn default() -> Self {
|
|
||||||
Self {
|
|
||||||
address: "0.0.0.0".to_owned(),
|
|
||||||
port: 8080,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
|
||||||
#[serde(default)]
|
|
||||||
pub struct CompressionCfg {
|
|
||||||
/// Enable gzip compression
|
|
||||||
pub gzip_en: bool,
|
|
||||||
/// Gzip compression level (0-9)
|
|
||||||
pub gzip_level: u8,
|
|
||||||
/// Enable brotli compression
|
|
||||||
pub brotli_en: bool,
|
|
||||||
/// Brozli compression level (0-11)
|
|
||||||
pub brotli_level: u8,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl Default for CompressionCfg {
|
|
||||||
fn default() -> Self {
|
|
||||||
Self {
|
|
||||||
gzip_en: false,
|
|
||||||
gzip_level: 6,
|
|
||||||
brotli_en: false,
|
|
||||||
brotli_level: 7,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl CompressionCfg {
|
|
||||||
pub fn enabled(&self) -> bool {
|
|
||||||
self.gzip_en || self.brotli_en
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Debug, Default, Clone, Serialize, Deserialize)]
|
|
||||||
#[serde(default)]
|
|
||||||
pub struct KeyCfg {
|
|
||||||
#[serde(skip_serializing_if = "Domains::is_none")]
|
|
||||||
pub domains: Domains,
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Debug, Default, Clone, Serialize, Deserialize)]
|
|
||||||
#[serde(untagged)]
|
|
||||||
pub enum Domains {
|
|
||||||
#[default]
|
|
||||||
None,
|
|
||||||
Single(String),
|
|
||||||
Multiple(Vec<String>),
|
|
||||||
}
|
|
||||||
|
|
||||||
impl Domains {
|
|
||||||
fn is_none(&self) -> bool {
|
|
||||||
matches!(self, Domains::None)
|
|
||||||
}
|
|
||||||
|
|
||||||
fn pattern_matches_domain(pattern: &str, domain: &str) -> bool {
|
|
||||||
if pattern == "*" {
|
|
||||||
true
|
|
||||||
} else if pattern.starts_with('/') && pattern.ends_with('/') {
|
|
||||||
let regex_str = &pattern[1..pattern.len() - 1];
|
|
||||||
let re = match Regex::new(regex_str) {
|
|
||||||
Ok(re) => re,
|
|
||||||
Err(e) => {
|
|
||||||
log::error!("could not parse regex `{regex_str}`, error: {e}");
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
};
|
|
||||||
re.is_match(domain)
|
|
||||||
} else {
|
|
||||||
domain == pattern
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn matches_domain(&self, domain: &str) -> bool {
|
|
||||||
match self {
|
|
||||||
Domains::None => false,
|
|
||||||
Domains::Single(pattern) => Self::pattern_matches_domain(pattern, domain),
|
|
||||||
Domains::Multiple(patterns) => patterns
|
|
||||||
.iter()
|
|
||||||
.any(|pattern| Self::pattern_matches_domain(pattern, domain)),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
#[cfg(test)]
|
|
||||||
mod tests {
|
|
||||||
use super::*;
|
|
||||||
|
|
||||||
use rstest::rstest;
|
|
||||||
|
|
||||||
#[rstest]
|
|
||||||
#[case("*", "hello-world", true)]
|
|
||||||
#[case("hello-world", "hello-world", true)]
|
|
||||||
#[case("hello-world", "hello-world2", false)]
|
|
||||||
#[case("/^talon-\\d+/", "talon-1", true)]
|
|
||||||
#[case("/^talon-\\d+/", "talon-x", false)]
|
|
||||||
fn pattern_matches_domain(#[case] pattern: &str, #[case] domain: &str, #[case] expect: bool) {
|
|
||||||
assert_eq!(Domains::pattern_matches_domain(pattern, domain), expect);
|
|
||||||
}
|
|
||||||
}
|
|
|
@ -378,15 +378,11 @@ impl Db {
|
||||||
}
|
}
|
||||||
|
|
||||||
fn file_key(version: u32, path: &str) -> String {
|
fn file_key(version: u32, path: &str) -> String {
|
||||||
|
// Remove leading/trailing slashes from path
|
||||||
|
let path = path.trim_matches('/');
|
||||||
format!("{version}:{path}")
|
format!("{version}:{path}")
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Get the hash of a file in the database
|
|
||||||
pub fn get_file_opt(&self, version: u32, path: &str) -> Result<Option<Vec<u8>>> {
|
|
||||||
let key = Self::file_key(version, path);
|
|
||||||
Ok(self.i.files.get(key)?.map(|hash| hash.to_vec()))
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Get the hash of a file in the database
|
/// Get the hash of a file in the database
|
||||||
pub fn get_file(&self, version: u32, path: &str) -> Result<Vec<u8>> {
|
pub fn get_file(&self, version: u32, path: &str) -> Result<Vec<u8>> {
|
||||||
let key = Self::file_key(version, path);
|
let key = Self::file_key(version, path);
|
||||||
|
|
|
@ -60,6 +60,8 @@ pub struct WebsiteUpdate {
|
||||||
pub source_url: Option<Option<String>>,
|
pub source_url: Option<Option<String>>,
|
||||||
/// Icon for the source link
|
/// Icon for the source link
|
||||||
pub source_icon: Option<Option<SourceIcon>>,
|
pub source_icon: Option<Option<SourceIcon>>,
|
||||||
|
/// File hash of the fallback page (for single page applications)
|
||||||
|
pub spa_fallback: Option<Option<String>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Website version stored in the database
|
/// Website version stored in the database
|
||||||
|
|
|
@ -1,4 +1,3 @@
|
||||||
pub mod config;
|
|
||||||
pub mod db;
|
pub mod db;
|
||||||
pub mod model;
|
pub mod model;
|
||||||
pub mod storage;
|
pub mod storage;
|
||||||
|
|
175
src/storage.rs
175
src/storage.rs
|
@ -1,21 +1,17 @@
|
||||||
use std::{
|
use std::{
|
||||||
borrow::Cow,
|
borrow::Cow,
|
||||||
collections::BTreeMap,
|
|
||||||
fs,
|
fs,
|
||||||
io::{BufReader, Read, Seek},
|
io::{Read, Seek},
|
||||||
path::{Path, PathBuf},
|
path::{Path, PathBuf},
|
||||||
sync::Arc,
|
sync::Arc,
|
||||||
};
|
};
|
||||||
|
|
||||||
use flate2::{read::GzDecoder, write::GzEncoder};
|
use flate2::read::GzDecoder;
|
||||||
use hex::ToHex;
|
use hex::ToHex;
|
||||||
use mime_guess::Mime;
|
|
||||||
use poem::http::HeaderMap;
|
|
||||||
use temp_dir::TempDir;
|
use temp_dir::TempDir;
|
||||||
use zip::ZipArchive;
|
use zip::ZipArchive;
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
config::Config,
|
|
||||||
db::{Db, DbError},
|
db::{Db, DbError},
|
||||||
util,
|
util,
|
||||||
};
|
};
|
||||||
|
@ -27,38 +23,6 @@ pub struct Storage {
|
||||||
struct StorageInner {
|
struct StorageInner {
|
||||||
path: PathBuf,
|
path: PathBuf,
|
||||||
db: Db,
|
db: Db,
|
||||||
cfg: Config,
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Debug, Default, Clone, Copy, PartialEq, Eq, PartialOrd, Ord)]
|
|
||||||
pub enum CompressionAlg {
|
|
||||||
#[default]
|
|
||||||
None,
|
|
||||||
Gzip,
|
|
||||||
Brotli,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl CompressionAlg {
|
|
||||||
/// Get value of the http encoding header
|
|
||||||
fn encoding(&self) -> Option<&'static str> {
|
|
||||||
match self {
|
|
||||||
CompressionAlg::None => None,
|
|
||||||
CompressionAlg::Gzip => Some("gzip"),
|
|
||||||
CompressionAlg::Brotli => Some("br"),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Debug)]
|
|
||||||
pub struct GotFile {
|
|
||||||
/// File path on disk
|
|
||||||
pub file_path: PathBuf,
|
|
||||||
/// File encoding
|
|
||||||
pub encoding: Option<&'static str>,
|
|
||||||
/// MIME type
|
|
||||||
pub mime: Option<Mime>,
|
|
||||||
/// Website path to redirect to
|
|
||||||
pub rd_path: Option<String>,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(thiserror::Error, Debug)]
|
#[derive(thiserror::Error, Debug)]
|
||||||
|
@ -71,10 +35,6 @@ pub enum StorageError {
|
||||||
InvalidFile(PathBuf),
|
InvalidFile(PathBuf),
|
||||||
#[error("zip archive error: {0}")]
|
#[error("zip archive error: {0}")]
|
||||||
Zip(#[from] zip::result::ZipError),
|
Zip(#[from] zip::result::ZipError),
|
||||||
#[error("page {0} not found")]
|
|
||||||
NotFound(String),
|
|
||||||
#[error("file {0} of page {1} missing from storage")]
|
|
||||||
MissingFile(String, String),
|
|
||||||
}
|
}
|
||||||
|
|
||||||
type Result<T> = std::result::Result<T, StorageError>;
|
type Result<T> = std::result::Result<T, StorageError>;
|
||||||
|
@ -83,12 +43,11 @@ const TMPDIR_PREFIX: &str = "talon";
|
||||||
|
|
||||||
impl Storage {
|
impl Storage {
|
||||||
/// Create a new file storage using the root folder and the database
|
/// Create a new file storage using the root folder and the database
|
||||||
pub fn new<P: Into<PathBuf>>(path: P, db: Db, cfg: Config) -> Self {
|
pub fn new<P: Into<PathBuf>>(path: P, db: Db) -> Self {
|
||||||
Self {
|
Self {
|
||||||
i: StorageInner {
|
i: StorageInner {
|
||||||
path: path.into(),
|
path: path.into(),
|
||||||
db,
|
db,
|
||||||
cfg,
|
|
||||||
}
|
}
|
||||||
.into(),
|
.into(),
|
||||||
}
|
}
|
||||||
|
@ -109,36 +68,13 @@ impl Storage {
|
||||||
let file_path = file_path.as_ref();
|
let file_path = file_path.as_ref();
|
||||||
|
|
||||||
let hash = util::hash_file(file_path)?;
|
let hash = util::hash_file(file_path)?;
|
||||||
let stored_file = self.file_path_mkdir(&hash)?;
|
let hash_str = hash.encode_hex::<String>();
|
||||||
|
|
||||||
fs::copy(file_path, &stored_file)?;
|
let subdir = self.i.path.join(&hash_str[..2]);
|
||||||
|
if !subdir.is_dir() {
|
||||||
if self.i.cfg.compression.enabled()
|
fs::create_dir(&subdir)?;
|
||||||
&& mime_guess::from_path(file_path)
|
|
||||||
.first()
|
|
||||||
.map(|t| compressible::is_compressible(t.essence_str()))
|
|
||||||
.unwrap_or_default()
|
|
||||||
{
|
|
||||||
if self.i.cfg.compression.gzip_en {
|
|
||||||
let mut encoder = GzEncoder::new(
|
|
||||||
fs::File::create(stored_file.with_extension("gz"))?,
|
|
||||||
flate2::Compression::new(self.i.cfg.compression.gzip_level.into()),
|
|
||||||
);
|
|
||||||
let mut input = BufReader::new(fs::File::open(&stored_file)?);
|
|
||||||
std::io::copy(&mut input, &mut encoder)?;
|
|
||||||
}
|
|
||||||
|
|
||||||
if self.i.cfg.compression.brotli_en {
|
|
||||||
let mut encoder = brotli::CompressorWriter::new(
|
|
||||||
fs::File::create(stored_file.with_extension("br"))?,
|
|
||||||
4096,
|
|
||||||
self.i.cfg.compression.brotli_level.into(),
|
|
||||||
20,
|
|
||||||
);
|
|
||||||
let mut input = BufReader::new(fs::File::open(&stored_file)?);
|
|
||||||
std::io::copy(&mut input, &mut encoder)?;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
fs::copy(file_path, subdir.join(&hash_str))?;
|
||||||
|
|
||||||
self.i.db.insert_file(version, site_path, &hash)?;
|
self.i.db.insert_file(version, site_path, &hash)?;
|
||||||
|
|
||||||
|
@ -222,99 +158,4 @@ impl Storage {
|
||||||
let import_path = Self::fix_archive_path(temp.path())?;
|
let import_path = Self::fix_archive_path(temp.path())?;
|
||||||
self.insert_dir(import_path, version)
|
self.insert_dir(import_path, version)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn file_path_mkdir(&self, hash: &[u8]) -> Result<PathBuf> {
|
|
||||||
let hash_str = hash.encode_hex::<String>();
|
|
||||||
|
|
||||||
let subdir = self.i.path.join(&hash_str[..2]);
|
|
||||||
if !subdir.is_dir() {
|
|
||||||
fs::create_dir(&subdir)?;
|
|
||||||
}
|
|
||||||
Ok(subdir.join(&hash_str))
|
|
||||||
}
|
|
||||||
|
|
||||||
fn file_path(&self, hash: &[u8]) -> PathBuf {
|
|
||||||
let hash_str = hash.encode_hex::<String>();
|
|
||||||
let subdir = self.i.path.join(&hash_str[..2]);
|
|
||||||
subdir.join(&hash_str)
|
|
||||||
}
|
|
||||||
|
|
||||||
fn files_compressed(&self, hash: &[u8]) -> BTreeMap<CompressionAlg, PathBuf> {
|
|
||||||
let path = self.file_path(hash);
|
|
||||||
let mut res = BTreeMap::new();
|
|
||||||
|
|
||||||
if self.i.cfg.compression.gzip_en {
|
|
||||||
let path_gz = path.with_extension("gz");
|
|
||||||
if path_gz.is_file() {
|
|
||||||
res.insert(CompressionAlg::Gzip, path_gz);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if self.i.cfg.compression.brotli_en {
|
|
||||||
let path_br = path.with_extension("br");
|
|
||||||
if path_br.is_file() {
|
|
||||||
res.insert(CompressionAlg::Brotli, path_br);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if path.is_file() {
|
|
||||||
res.insert(CompressionAlg::None, path);
|
|
||||||
}
|
|
||||||
res
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Get a file using the raw site path and the website version
|
|
||||||
///
|
|
||||||
/// HTTP headers are used to determine if the compressed version of a file should be returned.
|
|
||||||
pub fn get_file(&self, version: u32, site_path: &str, headers: &HeaderMap) -> Result<GotFile> {
|
|
||||||
let site_path = util::trim_site_path(site_path);
|
|
||||||
let mut new_path: Cow<str> = site_path.into();
|
|
||||||
let mut rd_path = None;
|
|
||||||
let mut hash = None;
|
|
||||||
|
|
||||||
if site_path.is_empty() {
|
|
||||||
// Special case, open index.html directly
|
|
||||||
new_path = "index.html".into();
|
|
||||||
} else {
|
|
||||||
// Attempt to access the following pages
|
|
||||||
// 1. Site path directly
|
|
||||||
// 2. Site path + `/index.html`
|
|
||||||
match self.i.db.get_file_opt(version, site_path)? {
|
|
||||||
Some(h) => {
|
|
||||||
hash = Some(h);
|
|
||||||
}
|
|
||||||
None => {
|
|
||||||
if util::site_path_ext(site_path).is_none() {
|
|
||||||
new_path = format!("{site_path}/index.html").into();
|
|
||||||
rd_path = Some(format!("{site_path}/"));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
let hash = match hash {
|
|
||||||
Some(hash) => hash,
|
|
||||||
None => self
|
|
||||||
.i
|
|
||||||
.db
|
|
||||||
.get_file_opt(version, &new_path)?
|
|
||||||
.ok_or_else(|| StorageError::NotFound(site_path.to_owned()))?,
|
|
||||||
};
|
|
||||||
|
|
||||||
let mime = util::site_path_mime(&new_path);
|
|
||||||
|
|
||||||
let files = self.files_compressed(&hash);
|
|
||||||
let file = util::parse_accept_encoding(headers, &files);
|
|
||||||
|
|
||||||
match file {
|
|
||||||
Some((compression, file)) => Ok(GotFile {
|
|
||||||
file_path: file.to_owned(),
|
|
||||||
encoding: compression.encoding(),
|
|
||||||
mime,
|
|
||||||
rd_path,
|
|
||||||
}),
|
|
||||||
None => Err(StorageError::MissingFile(
|
|
||||||
hash.encode_hex::<String>(),
|
|
||||||
new_path.into(),
|
|
||||||
)),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
133
src/util.rs
133
src/util.rs
|
@ -1,11 +1,7 @@
|
||||||
use std::{collections::BTreeMap, fs::File, path::Path, str::FromStr};
|
use std::{fs::File, path::Path};
|
||||||
|
|
||||||
use mime_guess::Mime;
|
|
||||||
use poem::http::{header, HeaderMap};
|
|
||||||
use sha2::{Digest, Sha256};
|
use sha2::{Digest, Sha256};
|
||||||
|
|
||||||
use crate::storage::CompressionAlg;
|
|
||||||
|
|
||||||
/// Get SHA256 hash of file
|
/// Get SHA256 hash of file
|
||||||
pub fn hash_file<P: AsRef<Path>>(file: P) -> Result<[u8; 32], std::io::Error> {
|
pub fn hash_file<P: AsRef<Path>>(file: P) -> Result<[u8; 32], std::io::Error> {
|
||||||
let mut hasher = Sha256::new();
|
let mut hasher = Sha256::new();
|
||||||
|
@ -13,130 +9,3 @@ pub fn hash_file<P: AsRef<Path>>(file: P) -> Result<[u8; 32], std::io::Error> {
|
||||||
std::io::copy(&mut file, &mut hasher)?;
|
std::io::copy(&mut file, &mut hasher)?;
|
||||||
Ok(hasher.finalize().into())
|
Ok(hasher.finalize().into())
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Return the file extension of a website path
|
|
||||||
pub fn site_path_ext(path: &str) -> Option<&str> {
|
|
||||||
let mut parts = path.split('.').rev();
|
|
||||||
parts
|
|
||||||
.next()
|
|
||||||
.filter(|ext| !ext.contains('/') && parts.next().is_some())
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn trim_site_path(path: &str) -> &str {
|
|
||||||
path.trim_matches('/')
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Get the file extension of a website path
|
|
||||||
pub fn site_path_mime(path: &str) -> Option<Mime> {
|
|
||||||
site_path_ext(path).and_then(|ext| mime_guess::from_ext(ext).first())
|
|
||||||
}
|
|
||||||
|
|
||||||
enum ContentCoding {
|
|
||||||
Brotli,
|
|
||||||
Gzip,
|
|
||||||
Star,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl FromStr for ContentCoding {
|
|
||||||
type Err = ();
|
|
||||||
|
|
||||||
fn from_str(s: &str) -> Result<Self, Self::Err> {
|
|
||||||
if s.eq_ignore_ascii_case("gzip") {
|
|
||||||
Ok(ContentCoding::Gzip)
|
|
||||||
} else if s.eq_ignore_ascii_case("br") {
|
|
||||||
Ok(ContentCoding::Brotli)
|
|
||||||
} else if s == "*" {
|
|
||||||
Ok(ContentCoding::Star)
|
|
||||||
} else {
|
|
||||||
Err(())
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Parse Accept-Encoding header and return the compressed file with the preferred algorithm
|
|
||||||
///
|
|
||||||
/// Source: <https://github.com/poem-web/poem/blob/049215cf02c5d4b1ab76f290b4708f3142d6d61b/poem/src/middleware/compression.rs#L36>
|
|
||||||
pub fn parse_accept_encoding<'a, T>(
|
|
||||||
headers: &HeaderMap,
|
|
||||||
files: &'a BTreeMap<CompressionAlg, T>,
|
|
||||||
) -> Option<(CompressionAlg, &'a T)> {
|
|
||||||
if files.is_empty() {
|
|
||||||
return None;
|
|
||||||
}
|
|
||||||
|
|
||||||
headers
|
|
||||||
.get_all(header::ACCEPT_ENCODING)
|
|
||||||
.iter()
|
|
||||||
.filter_map(|hval| hval.to_str().ok())
|
|
||||||
.flat_map(|s| s.split(',').map(str::trim))
|
|
||||||
.filter_map(|v| {
|
|
||||||
let (e, q) = match v.split_once(";q=") {
|
|
||||||
Some((e, q)) => (e, (q.parse::<f32>().ok()? * 1000.0) as i32),
|
|
||||||
None => (v, 1000),
|
|
||||||
};
|
|
||||||
let coding: ContentCoding = e.parse().ok()?;
|
|
||||||
let alg_file = match coding {
|
|
||||||
ContentCoding::Brotli => {
|
|
||||||
(CompressionAlg::Brotli, files.get(&CompressionAlg::Brotli)?)
|
|
||||||
}
|
|
||||||
ContentCoding::Gzip => (CompressionAlg::Gzip, files.get(&CompressionAlg::Gzip)?),
|
|
||||||
ContentCoding::Star => {
|
|
||||||
files.iter().max_by_key(|(a, _)| *a).map(|(a, f)| (*a, f))?
|
|
||||||
}
|
|
||||||
};
|
|
||||||
Some((alg_file, q))
|
|
||||||
})
|
|
||||||
.max_by_key(|((a, _), q)| (*q, *a))
|
|
||||||
.map(|(x, _)| x)
|
|
||||||
.or_else(|| {
|
|
||||||
files
|
|
||||||
.get(&CompressionAlg::None)
|
|
||||||
.map(|f| (CompressionAlg::None, f))
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
#[cfg(test)]
|
|
||||||
mod tests {
|
|
||||||
use super::*;
|
|
||||||
|
|
||||||
use rstest::rstest;
|
|
||||||
|
|
||||||
#[rstest]
|
|
||||||
#[case::html("index.html", Some("html"))]
|
|
||||||
#[case::none("hello.world/test", None)]
|
|
||||||
#[case::none("hello", None)]
|
|
||||||
fn t_site_path_ext(#[case] path: &str, #[case] expect: Option<&str>) {
|
|
||||||
let ext = site_path_ext(path);
|
|
||||||
assert_eq!(ext, expect)
|
|
||||||
}
|
|
||||||
|
|
||||||
#[rstest]
|
|
||||||
#[case::html("index.html", Some("text/html"))]
|
|
||||||
#[case::none("hello.world/test", None)]
|
|
||||||
fn t_site_path_mime(#[case] path: &str, #[case] expect: Option<&str>) {
|
|
||||||
let mime = site_path_mime(path).map(|mime| mime.essence_str().to_owned());
|
|
||||||
assert_eq!(mime.as_deref(), expect)
|
|
||||||
}
|
|
||||||
|
|
||||||
#[rstest]
|
|
||||||
#[case::none("", CompressionAlg::None)]
|
|
||||||
#[case::foo("foo", CompressionAlg::None)]
|
|
||||||
#[case::gz("gzip", CompressionAlg::Gzip)]
|
|
||||||
#[case::br("br", CompressionAlg::Brotli)]
|
|
||||||
#[case::star("*", CompressionAlg::Brotli)]
|
|
||||||
#[case::gz_deflate_br("gzip, deflate, br", CompressionAlg::Brotli)]
|
|
||||||
#[case::preference("br;q=0.8, gzip;q=1.0, *;q=0.1", CompressionAlg::Gzip)]
|
|
||||||
fn t_parse_accept_encoding(#[case] accept: &str, #[case] expect: CompressionAlg) {
|
|
||||||
let mut headers = HeaderMap::new();
|
|
||||||
headers.insert(header::ACCEPT_ENCODING, accept.parse().unwrap());
|
|
||||||
|
|
||||||
let mut files = BTreeMap::new();
|
|
||||||
files.insert(CompressionAlg::None, 0);
|
|
||||||
files.insert(CompressionAlg::Gzip, 1);
|
|
||||||
files.insert(CompressionAlg::Brotli, 2);
|
|
||||||
|
|
||||||
let (compression, file) = parse_accept_encoding(&headers, &files).unwrap();
|
|
||||||
assert_eq!(compression, expect);
|
|
||||||
assert_eq!(file, files.get(&compression).unwrap());
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
102
tests/fixtures/mod.rs
vendored
102
tests/fixtures/mod.rs
vendored
|
@ -1,18 +1,13 @@
|
||||||
use std::{collections::BTreeMap, ops::Deref};
|
use std::{collections::BTreeMap, ops::Deref};
|
||||||
|
|
||||||
use hex_literal::hex;
|
use hex_literal::hex;
|
||||||
use path_macro::path;
|
|
||||||
use rstest::fixture;
|
use rstest::fixture;
|
||||||
use temp_testdir::TempDir;
|
use temp_testdir::TempDir;
|
||||||
use time::macros::datetime;
|
use time::macros::datetime;
|
||||||
|
|
||||||
use talon::{
|
use talon::db::{
|
||||||
config::{CompressionCfg, Config, ConfigInner},
|
model::{Version, Website},
|
||||||
db::{
|
Db,
|
||||||
model::{Version, Website},
|
|
||||||
Db,
|
|
||||||
},
|
|
||||||
storage::Storage,
|
|
||||||
};
|
};
|
||||||
|
|
||||||
pub const SUBDOMAIN_1: &str = "";
|
pub const SUBDOMAIN_1: &str = "";
|
||||||
|
@ -29,12 +24,13 @@ pub const HASH_1_1_INDEX: [u8; 32] =
|
||||||
pub const HASH_1_1_STYLE: [u8; 32] =
|
pub const HASH_1_1_STYLE: [u8; 32] =
|
||||||
hex!("356f131c825fbf604797c7e9c85352549d81db8af91fee834016d075110af026");
|
hex!("356f131c825fbf604797c7e9c85352549d81db8af91fee834016d075110af026");
|
||||||
|
|
||||||
pub struct DbTest {
|
pub struct DbWrap {
|
||||||
db: Db,
|
db: Db,
|
||||||
_temp: TempDir,
|
#[allow(dead_code)]
|
||||||
|
temp: TempDir,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Deref for DbTest {
|
impl Deref for DbWrap {
|
||||||
type Target = Db;
|
type Target = Db;
|
||||||
|
|
||||||
fn deref(&self) -> &Self::Target {
|
fn deref(&self) -> &Self::Target {
|
||||||
|
@ -43,14 +39,34 @@ impl Deref for DbTest {
|
||||||
}
|
}
|
||||||
|
|
||||||
#[fixture]
|
#[fixture]
|
||||||
pub fn db_empty() -> DbTest {
|
pub fn db_sp() -> DbWrap {
|
||||||
let temp = temp_testdir::TempDir::default();
|
let temp = temp_testdir::TempDir::default();
|
||||||
let db = Db::new(&temp).unwrap();
|
let db = Db::new(&temp).unwrap();
|
||||||
|
|
||||||
DbTest { db, _temp: temp }
|
db.insert_website(
|
||||||
|
SUBDOMAIN_1,
|
||||||
|
&Website {
|
||||||
|
name: "ThetaDev".to_owned(),
|
||||||
|
created_at: datetime!(2023-02-18 16:30 +0),
|
||||||
|
latest_version: Some(VERSION_1_2),
|
||||||
|
icon: Some(
|
||||||
|
hex!("9f7e7971b4bfdb75429e534dea461ed90340886925078cda252cada9aa0e25f7").to_vec(),
|
||||||
|
),
|
||||||
|
color: Some(2068974),
|
||||||
|
visibility: talon::model::Visibility::Featured,
|
||||||
|
..Default::default()
|
||||||
|
},
|
||||||
|
)
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
DbWrap { db, temp }
|
||||||
}
|
}
|
||||||
|
|
||||||
fn insert_websites(db: &Db) {
|
#[fixture]
|
||||||
|
pub fn db() -> DbWrap {
|
||||||
|
let temp = temp_testdir::TempDir::default();
|
||||||
|
let db = Db::new(&temp).unwrap();
|
||||||
|
|
||||||
db.insert_website(
|
db.insert_website(
|
||||||
SUBDOMAIN_1,
|
SUBDOMAIN_1,
|
||||||
&Website {
|
&Website {
|
||||||
|
@ -153,13 +169,6 @@ fn insert_websites(db: &Db) {
|
||||||
},
|
},
|
||||||
)
|
)
|
||||||
.unwrap();
|
.unwrap();
|
||||||
}
|
|
||||||
|
|
||||||
#[fixture]
|
|
||||||
pub fn db() -> DbTest {
|
|
||||||
let temp = temp_testdir::TempDir::default();
|
|
||||||
let db = Db::new(&temp).unwrap();
|
|
||||||
insert_websites(&db);
|
|
||||||
|
|
||||||
db.insert_file(VERSION_1_1, "index.html", &HASH_1_1_INDEX)
|
db.insert_file(VERSION_1_1, "index.html", &HASH_1_1_INDEX)
|
||||||
.unwrap();
|
.unwrap();
|
||||||
|
@ -223,54 +232,5 @@ pub fn db() -> DbTest {
|
||||||
)
|
)
|
||||||
.unwrap();
|
.unwrap();
|
||||||
|
|
||||||
DbTest { db, _temp: temp }
|
DbWrap { db, temp }
|
||||||
}
|
|
||||||
|
|
||||||
pub struct StorageTest {
|
|
||||||
store: Storage,
|
|
||||||
_temp: TempDir,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl Deref for StorageTest {
|
|
||||||
type Target = Storage;
|
|
||||||
|
|
||||||
fn deref(&self) -> &Self::Target {
|
|
||||||
&self.store
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
#[fixture]
|
|
||||||
pub fn store() -> StorageTest {
|
|
||||||
let temp = temp_testdir::TempDir::default();
|
|
||||||
let db_path = path!(temp / "db");
|
|
||||||
std::fs::create_dir(&db_path).unwrap();
|
|
||||||
|
|
||||||
let cfg = Config::new(ConfigInner {
|
|
||||||
compression: CompressionCfg {
|
|
||||||
gzip_en: true,
|
|
||||||
brotli_en: true,
|
|
||||||
..Default::default()
|
|
||||||
},
|
|
||||||
..Default::default()
|
|
||||||
});
|
|
||||||
|
|
||||||
let db = Db::new(&db_path).unwrap();
|
|
||||||
insert_websites(&db);
|
|
||||||
|
|
||||||
let store = Storage::new(temp.to_path_buf(), db, cfg);
|
|
||||||
|
|
||||||
store
|
|
||||||
.insert_dir(path!("tests" / "testfiles" / "ThetaDev0"), VERSION_1_1)
|
|
||||||
.unwrap();
|
|
||||||
store
|
|
||||||
.insert_dir(path!("tests" / "testfiles" / "ThetaDev1"), VERSION_1_2)
|
|
||||||
.unwrap();
|
|
||||||
store
|
|
||||||
.insert_dir(path!("tests" / "testfiles" / "GenderEx"), VERSION_2_1)
|
|
||||||
.unwrap();
|
|
||||||
store
|
|
||||||
.insert_dir(path!("tests" / "testfiles" / "RustyPipe"), VERSION_3_1)
|
|
||||||
.unwrap();
|
|
||||||
|
|
||||||
StorageTest { store, _temp: temp }
|
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,30 +0,0 @@
|
||||||
---
|
|
||||||
source: tests/tests.rs
|
|
||||||
expression: "&cfg"
|
|
||||||
---
|
|
||||||
ConfigInner(
|
|
||||||
server: ServerCfg(
|
|
||||||
address: "127.0.0.1",
|
|
||||||
port: 3000,
|
|
||||||
),
|
|
||||||
compression: CompressionCfg(
|
|
||||||
gzip_en: true,
|
|
||||||
gzip_level: 6,
|
|
||||||
brotli_en: true,
|
|
||||||
brotli_level: 7,
|
|
||||||
),
|
|
||||||
keys: {
|
|
||||||
"04e99561e3824f387a217d141d2a3b46375de6864afbedf9c9a2cc102bc946a4": KeyCfg(
|
|
||||||
domains: "/^talon-\\d+/",
|
|
||||||
),
|
|
||||||
"21bdac19ffd22870d561b1d55b35eddd9029497107edb7b926aa3e7856bb409b": KeyCfg(
|
|
||||||
domains: [
|
|
||||||
"spotify-gender-ex",
|
|
||||||
"rustypipe",
|
|
||||||
],
|
|
||||||
),
|
|
||||||
"c32ff286c8ac1c3102625badf38ffd251ae0c4a56079d8ba490f320af63f1f47": KeyCfg(
|
|
||||||
domains: "*",
|
|
||||||
),
|
|
||||||
},
|
|
||||||
)
|
|
|
@ -1,28 +0,0 @@
|
||||||
---
|
|
||||||
source: tests/tests.rs
|
|
||||||
expression: "&cfg"
|
|
||||||
---
|
|
||||||
ConfigInner(
|
|
||||||
server: ServerCfg(
|
|
||||||
address: "0.0.0.0",
|
|
||||||
port: 8080,
|
|
||||||
),
|
|
||||||
compression: CompressionCfg(
|
|
||||||
gzip_en: true,
|
|
||||||
gzip_level: 6,
|
|
||||||
brotli_en: false,
|
|
||||||
brotli_level: 7,
|
|
||||||
),
|
|
||||||
keys: {
|
|
||||||
"04e99561e3824f387a217d141d2a3b46375de6864afbedf9c9a2cc102bc946a4": KeyCfg(),
|
|
||||||
"21bdac19ffd22870d561b1d55b35eddd9029497107edb7b926aa3e7856bb409b": KeyCfg(
|
|
||||||
domains: [
|
|
||||||
"spotify-gender-ex",
|
|
||||||
"rustypipe",
|
|
||||||
],
|
|
||||||
),
|
|
||||||
"c32ff286c8ac1c3102625badf38ffd251ae0c4a56079d8ba490f320af63f1f47": KeyCfg(
|
|
||||||
domains: "*",
|
|
||||||
),
|
|
||||||
},
|
|
||||||
)
|
|
|
@ -17,9 +17,6 @@
|
||||||
Client for the public YouTube / YouTube Music API (Innertube), inspired by
|
Client for the public YouTube / YouTube Music API (Innertube), inspired by
|
||||||
<a href="https://github.com/TeamNewPipe/NewPipeExtractor">NewPipe</a>.
|
<a href="https://github.com/TeamNewPipe/NewPipeExtractor">NewPipe</a>.
|
||||||
</p>
|
</p>
|
||||||
<p>
|
|
||||||
<a href="/page2">Page 2</a>
|
|
||||||
</p>
|
|
||||||
<p>
|
<p>
|
||||||
Carrot cake biscuit icing pudding danish topping powder. Croissant sugar plum
|
Carrot cake biscuit icing pudding danish topping powder. Croissant sugar plum
|
||||||
pudding halvah chocolate. Cotton candy tart cake bonbon tart. Shortbread jelly
|
pudding halvah chocolate. Cotton candy tart cake bonbon tart. Shortbread jelly
|
||||||
|
|
|
@ -1,109 +0,0 @@
|
||||||
<!DOCTYPE html>
|
|
||||||
<html lang="en">
|
|
||||||
<head>
|
|
||||||
<meta charset="UTF-8" />
|
|
||||||
<meta name="viewport" content="width=device-width, initial-scale=1.0" />
|
|
||||||
|
|
||||||
<meta name="title" content="Spotify-Gender-Ex" />
|
|
||||||
|
|
||||||
<link rel="stylesheet" type="text/css" href="/rp_style.css" />
|
|
||||||
|
|
||||||
<title>RustyPipe #2</title>
|
|
||||||
</head>
|
|
||||||
<body>
|
|
||||||
<div id="main">
|
|
||||||
<h1>RustyPipe #2</h1>
|
|
||||||
<p>
|
|
||||||
Client for the public YouTube / YouTube Music API (Innertube), inspired by
|
|
||||||
<a href="https://github.com/TeamNewPipe/NewPipeExtractor">NewPipe</a>.
|
|
||||||
</p>
|
|
||||||
<p>
|
|
||||||
<a href="/">Page 1</a>
|
|
||||||
</p>
|
|
||||||
<p>
|
|
||||||
Carrot cake biscuit icing pudding danish topping powder. Croissant sugar plum
|
|
||||||
pudding halvah chocolate. Cotton candy tart cake bonbon tart. Shortbread jelly
|
|
||||||
fruitcake icing pastry. Dragée dessert cupcake cake sesame snaps toffee pie.
|
|
||||||
Sweet roll sweet roll chupa chups jelly-o gummies tootsie roll sweet halvah oat
|
|
||||||
cake. Carrot cake carrot cake muffin bonbon sesame snaps brownie. Bonbon candy
|
|
||||||
macaroon fruitcake candy canes. Cake pudding danish liquorice cupcake jelly-o
|
|
||||||
ice cream. Liquorice lollipop danish tootsie roll toffee. Gingerbread chocolate
|
|
||||||
candy canes donut lemon drops apple pie danish bear claw. Caramels cake jelly
|
|
||||||
jelly sweet chocolate bar gingerbread icing. Cake soufflé lollipop pudding
|
|
||||||
marshmallow candy canes tootsie roll danish.
|
|
||||||
</p>
|
|
||||||
<p>
|
|
||||||
Carrot cake biscuit icing pudding danish topping powder. Croissant sugar plum
|
|
||||||
pudding halvah chocolate. Cotton candy tart cake bonbon tart. Shortbread jelly
|
|
||||||
fruitcake icing pastry. Dragée dessert cupcake cake sesame snaps toffee pie.
|
|
||||||
Sweet roll sweet roll chupa chups jelly-o gummies tootsie roll sweet halvah oat
|
|
||||||
cake. Carrot cake carrot cake muffin bonbon sesame snaps brownie. Bonbon candy
|
|
||||||
macaroon fruitcake candy canes. Cake pudding danish liquorice cupcake jelly-o
|
|
||||||
ice cream. Liquorice lollipop danish tootsie roll toffee. Gingerbread chocolate
|
|
||||||
candy canes donut lemon drops apple pie danish bear claw. Caramels cake jelly
|
|
||||||
jelly sweet chocolate bar gingerbread icing. Cake soufflé lollipop pudding
|
|
||||||
marshmallow candy canes tootsie roll danish.
|
|
||||||
</p>
|
|
||||||
<p>
|
|
||||||
Carrot cake biscuit icing pudding danish topping powder. Croissant sugar plum
|
|
||||||
pudding halvah chocolate. Cotton candy tart cake bonbon tart. Shortbread jelly
|
|
||||||
fruitcake icing pastry. Dragée dessert cupcake cake sesame snaps toffee pie.
|
|
||||||
Sweet roll sweet roll chupa chups jelly-o gummies tootsie roll sweet halvah oat
|
|
||||||
cake. Carrot cake carrot cake muffin bonbon sesame snaps brownie. Bonbon candy
|
|
||||||
macaroon fruitcake candy canes. Cake pudding danish liquorice cupcake jelly-o
|
|
||||||
ice cream. Liquorice lollipop danish tootsie roll toffee. Gingerbread chocolate
|
|
||||||
candy canes donut lemon drops apple pie danish bear claw. Caramels cake jelly
|
|
||||||
jelly sweet chocolate bar gingerbread icing. Cake soufflé lollipop pudding
|
|
||||||
marshmallow candy canes tootsie roll danish.
|
|
||||||
</p>
|
|
||||||
<p>
|
|
||||||
Carrot cake biscuit icing pudding danish topping powder. Croissant sugar plum
|
|
||||||
pudding halvah chocolate. Cotton candy tart cake bonbon tart. Shortbread jelly
|
|
||||||
fruitcake icing pastry. Dragée dessert cupcake cake sesame snaps toffee pie.
|
|
||||||
Sweet roll sweet roll chupa chups jelly-o gummies tootsie roll sweet halvah oat
|
|
||||||
cake. Carrot cake carrot cake muffin bonbon sesame snaps brownie. Bonbon candy
|
|
||||||
macaroon fruitcake candy canes. Cake pudding danish liquorice cupcake jelly-o
|
|
||||||
ice cream. Liquorice lollipop danish tootsie roll toffee. Gingerbread chocolate
|
|
||||||
candy canes donut lemon drops apple pie danish bear claw. Caramels cake jelly
|
|
||||||
jelly sweet chocolate bar gingerbread icing. Cake soufflé lollipop pudding
|
|
||||||
marshmallow candy canes tootsie roll danish.
|
|
||||||
</p>
|
|
||||||
<p>
|
|
||||||
Carrot cake biscuit icing pudding danish topping powder. Croissant sugar plum
|
|
||||||
pudding halvah chocolate. Cotton candy tart cake bonbon tart. Shortbread jelly
|
|
||||||
fruitcake icing pastry. Dragée dessert cupcake cake sesame snaps toffee pie.
|
|
||||||
Sweet roll sweet roll chupa chups jelly-o gummies tootsie roll sweet halvah oat
|
|
||||||
cake. Carrot cake carrot cake muffin bonbon sesame snaps brownie. Bonbon candy
|
|
||||||
macaroon fruitcake candy canes. Cake pudding danish liquorice cupcake jelly-o
|
|
||||||
ice cream. Liquorice lollipop danish tootsie roll toffee. Gingerbread chocolate
|
|
||||||
candy canes donut lemon drops apple pie danish bear claw. Caramels cake jelly
|
|
||||||
jelly sweet chocolate bar gingerbread icing. Cake soufflé lollipop pudding
|
|
||||||
marshmallow candy canes tootsie roll danish.
|
|
||||||
</p>
|
|
||||||
<p>
|
|
||||||
Carrot cake biscuit icing pudding danish topping powder. Croissant sugar plum
|
|
||||||
pudding halvah chocolate. Cotton candy tart cake bonbon tart. Shortbread jelly
|
|
||||||
fruitcake icing pastry. Dragée dessert cupcake cake sesame snaps toffee pie.
|
|
||||||
Sweet roll sweet roll chupa chups jelly-o gummies tootsie roll sweet halvah oat
|
|
||||||
cake. Carrot cake carrot cake muffin bonbon sesame snaps brownie. Bonbon candy
|
|
||||||
macaroon fruitcake candy canes. Cake pudding danish liquorice cupcake jelly-o
|
|
||||||
ice cream. Liquorice lollipop danish tootsie roll toffee. Gingerbread chocolate
|
|
||||||
candy canes donut lemon drops apple pie danish bear claw. Caramels cake jelly
|
|
||||||
jelly sweet chocolate bar gingerbread icing. Cake soufflé lollipop pudding
|
|
||||||
marshmallow candy canes tootsie roll danish.
|
|
||||||
</p>
|
|
||||||
<p>
|
|
||||||
Carrot cake biscuit icing pudding danish topping powder. Croissant sugar plum
|
|
||||||
pudding halvah chocolate. Cotton candy tart cake bonbon tart. Shortbread jelly
|
|
||||||
fruitcake icing pastry. Dragée dessert cupcake cake sesame snaps toffee pie.
|
|
||||||
Sweet roll sweet roll chupa chups jelly-o gummies tootsie roll sweet halvah oat
|
|
||||||
cake. Carrot cake carrot cake muffin bonbon sesame snaps brownie. Bonbon candy
|
|
||||||
macaroon fruitcake candy canes. Cake pudding danish liquorice cupcake jelly-o
|
|
||||||
ice cream. Liquorice lollipop danish tootsie roll toffee. Gingerbread chocolate
|
|
||||||
candy canes donut lemon drops apple pie danish bear claw. Caramels cake jelly
|
|
||||||
jelly sweet chocolate bar gingerbread icing. Cake soufflé lollipop pudding
|
|
||||||
marshmallow candy canes tootsie roll danish.
|
|
||||||
</p>
|
|
||||||
</div>
|
|
||||||
</body>
|
|
||||||
</html>
|
|
|
@ -1,23 +0,0 @@
|
||||||
[server]
|
|
||||||
address = "127.0.0.1"
|
|
||||||
port = 3000
|
|
||||||
|
|
||||||
# Talon compresses files when they are uploaded
|
|
||||||
# Here you can configure compression algorithms and levels
|
|
||||||
[compression]
|
|
||||||
gzip_en = true
|
|
||||||
gzip_level = 6
|
|
||||||
brotli_en = true
|
|
||||||
brotli_level = 7
|
|
||||||
|
|
||||||
# API keys for uploading websites
|
|
||||||
# You can configure the allowed domains per key (either a single string or a list of strings)
|
|
||||||
# Regexes can be used if they start and end with a slash
|
|
||||||
[keys.c32ff286c8ac1c3102625badf38ffd251ae0c4a56079d8ba490f320af63f1f47]
|
|
||||||
domains = "*"
|
|
||||||
|
|
||||||
[keys.21bdac19ffd22870d561b1d55b35eddd9029497107edb7b926aa3e7856bb409b]
|
|
||||||
domains = ["spotify-gender-ex", "rustypipe"]
|
|
||||||
|
|
||||||
[keys.04e99561e3824f387a217d141d2a3b46375de6864afbedf9c9a2cc102bc946a4]
|
|
||||||
domains = "/^talon-\\d+/"
|
|
|
@ -1,16 +0,0 @@
|
||||||
# Talon compresses files when they are uploaded
|
|
||||||
# Here you can configure compression algorithms and levels
|
|
||||||
[compression]
|
|
||||||
gzip_en = true
|
|
||||||
gzip_level = 6
|
|
||||||
|
|
||||||
# API keys for uploading websites
|
|
||||||
# You can configure the allowed domains per key (either a single string or a list of strings)
|
|
||||||
# Regexes can be used if they start and end with a slash
|
|
||||||
[keys.c32ff286c8ac1c3102625badf38ffd251ae0c4a56079d8ba490f320af63f1f47]
|
|
||||||
domains = "*"
|
|
||||||
|
|
||||||
[keys.21bdac19ffd22870d561b1d55b35eddd9029497107edb7b926aa3e7856bb409b]
|
|
||||||
domains = ["spotify-gender-ex", "rustypipe"]
|
|
||||||
|
|
||||||
[keys.04e99561e3824f387a217d141d2a3b46375de6864afbedf9c9a2cc102bc946a4]
|
|
148
tests/tests.rs
148
tests/tests.rs
|
@ -6,13 +6,12 @@ use path_macro::path;
|
||||||
use rstest::rstest;
|
use rstest::rstest;
|
||||||
|
|
||||||
use fixtures::*;
|
use fixtures::*;
|
||||||
use talon::db::{Db, DbError};
|
use talon::db::{model::WebsiteUpdate, Db, DbError};
|
||||||
|
use talon::storage::Storage;
|
||||||
|
|
||||||
mod database {
|
mod database {
|
||||||
use super::*;
|
use super::*;
|
||||||
|
|
||||||
use talon::db::model::WebsiteUpdate;
|
|
||||||
|
|
||||||
fn get_export(db: &Db) -> String {
|
fn get_export(db: &Db) -> String {
|
||||||
let mut buf: Vec<u8> = Vec::new();
|
let mut buf: Vec<u8> = Vec::new();
|
||||||
db.export(&mut buf).unwrap();
|
db.export(&mut buf).unwrap();
|
||||||
|
@ -20,13 +19,13 @@ mod database {
|
||||||
}
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn export(db: DbTest) {
|
fn export(db: DbWrap) {
|
||||||
let data = get_export(&db);
|
let data = get_export(&db);
|
||||||
insta::assert_snapshot!("export", data);
|
insta::assert_snapshot!("export", data);
|
||||||
}
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn export_import(db: DbTest) {
|
fn export_import(db: DbWrap) {
|
||||||
let td = temp_testdir::TempDir::default();
|
let td = temp_testdir::TempDir::default();
|
||||||
let p_export = td.join("export.jsonl");
|
let p_export = td.join("export.jsonl");
|
||||||
let p_db2 = td.join("db2");
|
let p_db2 = td.join("db2");
|
||||||
|
@ -42,7 +41,7 @@ mod database {
|
||||||
}
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn get_website(db: DbTest) {
|
fn get_website(db: DbWrap) {
|
||||||
let ws1 = db.get_website("").unwrap();
|
let ws1 = db.get_website("").unwrap();
|
||||||
let ws2 = db.get_website("spotify-gender-ex").unwrap();
|
let ws2 = db.get_website("spotify-gender-ex").unwrap();
|
||||||
let ws3 = db.get_website("rustypipe").unwrap();
|
let ws3 = db.get_website("rustypipe").unwrap();
|
||||||
|
@ -50,7 +49,7 @@ mod database {
|
||||||
}
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn delete_website(db: DbTest) {
|
fn delete_website(db: DbWrap) {
|
||||||
db.delete_website("", true).unwrap();
|
db.delete_website("", true).unwrap();
|
||||||
|
|
||||||
assert!(matches!(
|
assert!(matches!(
|
||||||
|
@ -68,7 +67,7 @@ mod database {
|
||||||
}
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn update_website(db: DbTest) {
|
fn update_website(db: DbWrap) {
|
||||||
db.update_website(
|
db.update_website(
|
||||||
"",
|
"",
|
||||||
WebsiteUpdate {
|
WebsiteUpdate {
|
||||||
|
@ -88,19 +87,19 @@ mod database {
|
||||||
}
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn get_websites(db: DbTest) {
|
fn get_websites(db: DbWrap) {
|
||||||
let websites = db.get_websites().map(|w| w.unwrap()).collect::<Vec<_>>();
|
let websites = db.get_websites().map(|w| w.unwrap()).collect::<Vec<_>>();
|
||||||
insta::assert_ron_snapshot!(websites);
|
insta::assert_ron_snapshot!(websites);
|
||||||
}
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn get_version(db: DbTest) {
|
fn get_version(db: DbWrap) {
|
||||||
let version = db.get_version("", VERSION_1_1).unwrap();
|
let version = db.get_version("", VERSION_1_1).unwrap();
|
||||||
insta::assert_ron_snapshot!(version);
|
insta::assert_ron_snapshot!(version);
|
||||||
}
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn delete_version(db: DbTest) {
|
fn delete_version(db: DbWrap) {
|
||||||
db.delete_version("", VERSION_1_2, true).unwrap();
|
db.delete_version("", VERSION_1_2, true).unwrap();
|
||||||
assert!(matches!(
|
assert!(matches!(
|
||||||
db.get_version("", VERSION_1_2).unwrap_err(),
|
db.get_version("", VERSION_1_2).unwrap_err(),
|
||||||
|
@ -121,7 +120,7 @@ mod database {
|
||||||
}
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn get_website_versions(db: DbTest) {
|
fn get_website_versions(db: DbWrap) {
|
||||||
let versions = db
|
let versions = db
|
||||||
.get_website_versions("")
|
.get_website_versions("")
|
||||||
.map(|v| v.unwrap())
|
.map(|v| v.unwrap())
|
||||||
|
@ -130,7 +129,7 @@ mod database {
|
||||||
}
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn get_website_version_ids(db: DbTest) {
|
fn get_website_version_ids(db: DbWrap) {
|
||||||
let ids = db
|
let ids = db
|
||||||
.get_website_version_ids("")
|
.get_website_version_ids("")
|
||||||
.map(|v| v.unwrap())
|
.map(|v| v.unwrap())
|
||||||
|
@ -139,13 +138,13 @@ mod database {
|
||||||
}
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn get_file(db: DbTest) {
|
fn get_file(db: DbWrap) {
|
||||||
let hash = db.get_file(VERSION_1_1, "index.html").unwrap();
|
let hash = db.get_file(VERSION_1_1, "index.html").unwrap();
|
||||||
assert_eq!(hash, HASH_1_1_INDEX);
|
assert_eq!(hash, HASH_1_1_INDEX);
|
||||||
}
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn delete_file(db: DbTest) {
|
fn delete_file(db: DbWrap) {
|
||||||
db.delete_file(VERSION_1_1, "index.html", true).unwrap();
|
db.delete_file(VERSION_1_1, "index.html", true).unwrap();
|
||||||
assert!(matches!(
|
assert!(matches!(
|
||||||
db.get_file(VERSION_1_1, "index.html").unwrap_err(),
|
db.get_file(VERSION_1_1, "index.html").unwrap_err(),
|
||||||
|
@ -159,7 +158,7 @@ mod database {
|
||||||
}
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn get_version_files(db: DbTest) {
|
fn get_version_files(db: DbWrap) {
|
||||||
let files = db
|
let files = db
|
||||||
.get_version_files(VERSION_1_1)
|
.get_version_files(VERSION_1_1)
|
||||||
.map(|f| f.unwrap())
|
.map(|f| f.unwrap())
|
||||||
|
@ -174,7 +173,7 @@ mod database {
|
||||||
}
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn get_file_hashes(db: DbTest) {
|
fn get_file_hashes(db: DbWrap) {
|
||||||
let hashes = db.get_file_hashes().unwrap();
|
let hashes = db.get_file_hashes().unwrap();
|
||||||
assert_eq!(hashes.len(), 12)
|
assert_eq!(hashes.len(), 12)
|
||||||
}
|
}
|
||||||
|
@ -182,21 +181,18 @@ mod database {
|
||||||
|
|
||||||
mod storage {
|
mod storage {
|
||||||
use hex::ToHex;
|
use hex::ToHex;
|
||||||
use poem::http::{header, HeaderMap};
|
|
||||||
use talon::config::{CompressionCfg, Config, ConfigInner};
|
|
||||||
use talon::storage::Storage;
|
|
||||||
|
|
||||||
use super::*;
|
use super::*;
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn insert_files(db_empty: DbTest) {
|
fn insert_files(db_sp: DbWrap) {
|
||||||
let dir = path!("tests" / "testfiles" / "ThetaDev1");
|
let dir = path!("tests" / "testfiles" / "ThetaDev1");
|
||||||
let temp = temp_testdir::TempDir::default();
|
let temp = temp_testdir::TempDir::default();
|
||||||
let store = Storage::new(temp.to_path_buf(), db_empty.clone(), Default::default());
|
let store = Storage::new(temp.to_path_buf(), db_sp.clone());
|
||||||
|
|
||||||
store.insert_dir(dir, 1).unwrap();
|
store.insert_dir(dir, 1).unwrap();
|
||||||
|
|
||||||
let files = db_empty
|
let files = db_sp
|
||||||
.get_version_files(1)
|
.get_version_files(1)
|
||||||
.map(|f| f.unwrap())
|
.map(|f| f.unwrap())
|
||||||
.collect::<Vec<_>>();
|
.collect::<Vec<_>>();
|
||||||
|
@ -210,16 +206,16 @@ mod storage {
|
||||||
}
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn insert_zip_archive(db_empty: DbTest) {
|
fn insert_zip_archive(db_sp: DbWrap) {
|
||||||
let archive = path!("tests" / "testfiles" / "archive" / "ThetaDev1.zip");
|
let archive = path!("tests" / "testfiles" / "archive" / "ThetaDev1.zip");
|
||||||
let temp = temp_testdir::TempDir::default();
|
let temp = temp_testdir::TempDir::default();
|
||||||
let store = Storage::new(temp.to_path_buf(), db_empty.clone(), Default::default());
|
let store = Storage::new(temp.to_path_buf(), db_sp.clone());
|
||||||
|
|
||||||
store
|
store
|
||||||
.insert_zip_archive(File::open(archive).unwrap(), 1)
|
.insert_zip_archive(File::open(archive).unwrap(), 1)
|
||||||
.unwrap();
|
.unwrap();
|
||||||
|
|
||||||
let files = db_empty
|
let files = db_sp
|
||||||
.get_version_files(1)
|
.get_version_files(1)
|
||||||
.map(|f| f.unwrap())
|
.map(|f| f.unwrap())
|
||||||
.collect::<Vec<_>>();
|
.collect::<Vec<_>>();
|
||||||
|
@ -233,16 +229,16 @@ mod storage {
|
||||||
}
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn insert_tgz_archive(db_empty: DbTest) {
|
fn insert_tgz_archive(db_sp: DbWrap) {
|
||||||
let archive = path!("tests" / "testfiles" / "archive" / "ThetaDev1.tar.gz");
|
let archive = path!("tests" / "testfiles" / "archive" / "ThetaDev1.tar.gz");
|
||||||
let temp = temp_testdir::TempDir::default();
|
let temp = temp_testdir::TempDir::default();
|
||||||
let store = Storage::new(temp.to_path_buf(), db_empty.clone(), Default::default());
|
let store = Storage::new(temp.to_path_buf(), db_sp.clone());
|
||||||
|
|
||||||
store
|
store
|
||||||
.insert_tgz_archive(File::open(archive).unwrap(), 1)
|
.insert_tgz_archive(File::open(archive).unwrap(), 1)
|
||||||
.unwrap();
|
.unwrap();
|
||||||
|
|
||||||
let files = db_empty
|
let files = db_sp
|
||||||
.get_version_files(1)
|
.get_version_files(1)
|
||||||
.map(|f| f.unwrap())
|
.map(|f| f.unwrap())
|
||||||
.collect::<Vec<_>>();
|
.collect::<Vec<_>>();
|
||||||
|
@ -254,98 +250,4 @@ mod storage {
|
||||||
assert!(path.is_file());
|
assert!(path.is_file());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[rstest]
|
|
||||||
#[case::gzip(CompressionCfg {gzip_en: true, ..Default::default()}, "gz")]
|
|
||||||
#[case::brotli(CompressionCfg {brotli_en: true, ..Default::default()}, "br")]
|
|
||||||
fn insert_files_compressed(
|
|
||||||
db_empty: DbTest,
|
|
||||||
#[case] compression: CompressionCfg,
|
|
||||||
#[case] ext: &str,
|
|
||||||
) {
|
|
||||||
let dir = path!("tests" / "testfiles" / "ThetaDev1");
|
|
||||||
let temp = temp_testdir::TempDir::default();
|
|
||||||
let cfg = Config::new(ConfigInner {
|
|
||||||
compression,
|
|
||||||
..Default::default()
|
|
||||||
});
|
|
||||||
|
|
||||||
let store = Storage::new(temp.to_path_buf(), db_empty.clone(), cfg);
|
|
||||||
store.insert_dir(dir, 1).unwrap();
|
|
||||||
|
|
||||||
for f in db_empty.get_version_files(1) {
|
|
||||||
let hash = f.unwrap().1;
|
|
||||||
let hash_str = hash.encode_hex::<String>();
|
|
||||||
let path = temp.join(&hash_str[..2]).join(&hash_str);
|
|
||||||
let path_compressed = path.with_extension(ext);
|
|
||||||
assert!(path.is_file());
|
|
||||||
|
|
||||||
// Images should not be compressed
|
|
||||||
let expect = &hash_str
|
|
||||||
!= "901d291a47a8a9b55c06f84e5e5f82fd2dcee65cac1406d6e878b805d45c1e93"
|
|
||||||
&& &hash_str != "9f7e7971b4bfdb75429e534dea461ed90340886925078cda252cada9aa0e25f7";
|
|
||||||
assert_eq!(path_compressed.is_file(), expect)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
#[rstest]
|
|
||||||
#[case::nocmp("", VERSION_1_2, "", true, "text/html", None)]
|
|
||||||
#[case::gzip("gzip", VERSION_1_2, "", true, "text/html", None)]
|
|
||||||
#[case::br("br", VERSION_1_2, "", true, "text/html", None)]
|
|
||||||
#[case::image("br", VERSION_1_2, "assets/image.jpg", false, "image/jpeg", None)]
|
|
||||||
#[case::subdir("br", VERSION_3_1, "page2", true, "text/html", Some("page2/"))]
|
|
||||||
fn get_file(
|
|
||||||
store: StorageTest,
|
|
||||||
#[case] encoding: &str,
|
|
||||||
#[case] version: u32,
|
|
||||||
#[case] path: &str,
|
|
||||||
#[case] compressible: bool,
|
|
||||||
#[case] mime: &str,
|
|
||||||
#[case] rd_path: Option<&str>,
|
|
||||||
) {
|
|
||||||
let mut headers = HeaderMap::new();
|
|
||||||
headers.insert(header::ACCEPT_ENCODING, encoding.parse().unwrap());
|
|
||||||
|
|
||||||
let expect_ext = if compressible {
|
|
||||||
match encoding {
|
|
||||||
"gzip" => Some("gz"),
|
|
||||||
"" => None,
|
|
||||||
e => Some(e),
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
None
|
|
||||||
};
|
|
||||||
|
|
||||||
let index_file = store.get_file(version, path, &headers).unwrap();
|
|
||||||
assert!(index_file.file_path.is_file());
|
|
||||||
assert_eq!(
|
|
||||||
index_file
|
|
||||||
.file_path
|
|
||||||
.extension()
|
|
||||||
.map(|s| s.to_str().unwrap()),
|
|
||||||
expect_ext
|
|
||||||
);
|
|
||||||
assert_eq!(
|
|
||||||
index_file.encoding,
|
|
||||||
Some(encoding).filter(|s| compressible && !s.is_empty())
|
|
||||||
);
|
|
||||||
assert_eq!(index_file.mime.unwrap().essence_str(), mime);
|
|
||||||
assert_eq!(index_file.rd_path.as_deref(), rd_path);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
mod config {
|
|
||||||
use talon::config::Config;
|
|
||||||
|
|
||||||
use super::*;
|
|
||||||
|
|
||||||
#[rstest]
|
|
||||||
#[case::default("default", "config.toml")]
|
|
||||||
#[case::sparse("sparse", "config_sparse.toml")]
|
|
||||||
fn parse_config(#[case] name: &str, #[case] fname: &str) {
|
|
||||||
let p = path!("tests" / "testfiles" / "config" / fname);
|
|
||||||
let cfg = Config::from_file(p).unwrap();
|
|
||||||
|
|
||||||
insta::assert_ron_snapshot!(name, &cfg);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in a new issue