Compare commits

...

6 commits

14 changed files with 374 additions and 257 deletions

160
Cargo.lock generated
View file

@ -238,6 +238,15 @@ dependencies = [
"memchr", "memchr",
] ]
[[package]]
name = "ansi_term"
version = "0.12.1"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "d52a9bb7ec0cf484c551830a7ce27bd20d67eac647e1befb56b0be4ee39a55d2"
dependencies = [
"winapi",
]
[[package]] [[package]]
name = "anyhow" name = "anyhow"
version = "1.0.42" version = "1.0.42"
@ -409,6 +418,17 @@ version = "1.0.0"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "baf1de4339761588bc0619e3cbc0120ee582ebb74b53b4efbf79117bd2da40fd" checksum = "baf1de4339761588bc0619e3cbc0120ee582ebb74b53b4efbf79117bd2da40fd"
[[package]]
name = "chacha20"
version = "0.7.1"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "fee7ad89dc1128635074c268ee661f90c3f7e83d9fd12910608c36b47d6c3412"
dependencies = [
"cfg-if",
"cipher",
"cpufeatures",
]
[[package]] [[package]]
name = "chrono" name = "chrono"
version = "0.4.19" version = "0.4.19"
@ -423,6 +443,15 @@ dependencies = [
"winapi", "winapi",
] ]
[[package]]
name = "cipher"
version = "0.3.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "7ee52072ec15386f770805afd189a01c8841be8696bed250fa2f13c4c0d6dfb7"
dependencies = [
"generic-array",
]
[[package]] [[package]]
name = "clap" name = "clap"
version = "3.0.0-beta.2" version = "3.0.0-beta.2"
@ -456,17 +485,6 @@ dependencies = [
"syn", "syn",
] ]
[[package]]
name = "colored"
version = "1.9.3"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "f4ffc801dacf156c5854b9df4f425a626539c3a6ef7893cc0c5084a23f0b6c59"
dependencies = [
"atty",
"lazy_static",
"winapi",
]
[[package]] [[package]]
name = "const_fn" name = "const_fn"
version = "0.4.8" version = "0.4.8"
@ -1179,6 +1197,7 @@ dependencies = [
"base64", "base64",
"bincode", "bincode",
"bytes", "bytes",
"chacha20",
"chrono", "chrono",
"clap", "clap",
"ctrlc", "ctrlc",
@ -1197,7 +1216,6 @@ dependencies = [
"serde_json", "serde_json",
"serde_repr", "serde_repr",
"serde_yaml", "serde_yaml",
"simple_logger",
"sodiumoxide", "sodiumoxide",
"sqlx", "sqlx",
"tempfile", "tempfile",
@ -1205,6 +1223,9 @@ dependencies = [
"tokio", "tokio",
"tokio-stream", "tokio-stream",
"tokio-util", "tokio-util",
"tracing",
"tracing-futures",
"tracing-subscriber",
"url", "url",
"vergen", "vergen",
] ]
@ -1215,6 +1236,15 @@ version = "1.0.2"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "3e2e65a1a2e43cfcb47a895c4c8b10d1f4a61097f9f254f183aee60cad9c651d" checksum = "3e2e65a1a2e43cfcb47a895c4c8b10d1f4a61097f9f254f183aee60cad9c651d"
[[package]]
name = "matchers"
version = "0.0.1"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "f099785f7595cc4b4553a174ce30dd7589ef93391ff414dbb67f62392b9e0ce1"
dependencies = [
"regex-automata",
]
[[package]] [[package]]
name = "matches" name = "matches"
version = "0.1.8" version = "0.1.8"
@ -1604,6 +1634,15 @@ dependencies = [
"regex-syntax", "regex-syntax",
] ]
[[package]]
name = "regex-automata"
version = "0.1.10"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "6c230d73fb8d8c1b9c0b3135c5142a8acee3a0558fb8db5cf1cb65f8d7862132"
dependencies = [
"regex-syntax",
]
[[package]] [[package]]
name = "regex-syntax" name = "regex-syntax"
version = "0.6.25" version = "0.6.25"
@ -1885,6 +1924,15 @@ dependencies = [
"opaque-debug", "opaque-debug",
] ]
[[package]]
name = "sharded-slab"
version = "0.1.1"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "79c719719ee05df97490f80a45acfc99e5a30ce98a1e4fb67aee422745ae14e3"
dependencies = [
"lazy_static",
]
[[package]] [[package]]
name = "signal-hook-registry" name = "signal-hook-registry"
version = "1.4.0" version = "1.4.0"
@ -1900,19 +1948,6 @@ version = "1.3.1"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "c19772be3c4dd2ceaacf03cb41d5885f2a02c4d8804884918e3a258480803335" checksum = "c19772be3c4dd2ceaacf03cb41d5885f2a02c4d8804884918e3a258480803335"
[[package]]
name = "simple_logger"
version = "1.11.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "cd57f17c093ead1d4a1499dc9acaafdd71240908d64775465543b8d9a9f1d198"
dependencies = [
"atty",
"chrono",
"colored",
"log",
"winapi",
]
[[package]] [[package]]
name = "slab" name = "slab"
version = "0.4.3" version = "0.4.3"
@ -2232,6 +2267,15 @@ dependencies = [
"syn", "syn",
] ]
[[package]]
name = "thread_local"
version = "1.1.3"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "8018d24e04c95ac8790716a5987d0fec4f8b27249ffa0f7d33f1369bdfb88cbd"
dependencies = [
"once_cell",
]
[[package]] [[package]]
name = "time" name = "time"
version = "0.1.43" version = "0.1.43"
@ -2377,9 +2421,21 @@ checksum = "09adeb8c97449311ccd28a427f96fb563e7fd31aabf994189879d9da2394b89d"
dependencies = [ dependencies = [
"cfg-if", "cfg-if",
"pin-project-lite", "pin-project-lite",
"tracing-attributes",
"tracing-core", "tracing-core",
] ]
[[package]]
name = "tracing-attributes"
version = "0.1.15"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "c42e6fa53307c8a17e4ccd4dc81cf5ec38db9209f59b222210375b54ee40d1e2"
dependencies = [
"proc-macro2",
"quote",
"syn",
]
[[package]] [[package]]
name = "tracing-core" name = "tracing-core"
version = "0.1.18" version = "0.1.18"
@ -2389,6 +2445,60 @@ dependencies = [
"lazy_static", "lazy_static",
] ]
[[package]]
name = "tracing-futures"
version = "0.2.5"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "97d095ae15e245a057c8e8451bab9b3ee1e1f68e9ba2b4fbc18d0ac5237835f2"
dependencies = [
"pin-project",
"tracing",
]
[[package]]
name = "tracing-log"
version = "0.1.2"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "a6923477a48e41c1951f1999ef8bb5a3023eb723ceadafe78ffb65dc366761e3"
dependencies = [
"lazy_static",
"log",
"tracing-core",
]
[[package]]
name = "tracing-serde"
version = "0.1.2"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "fb65ea441fbb84f9f6748fd496cf7f63ec9af5bca94dd86456978d055e8eb28b"
dependencies = [
"serde",
"tracing-core",
]
[[package]]
name = "tracing-subscriber"
version = "0.2.19"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "ab69019741fca4d98be3c62d2b75254528b5432233fd8a4d2739fec20278de48"
dependencies = [
"ansi_term",
"chrono",
"lazy_static",
"matchers",
"parking_lot",
"regex",
"serde",
"serde_json",
"sharded-slab",
"smallvec",
"thread_local",
"tracing",
"tracing-core",
"tracing-log",
"tracing-serde",
]
[[package]] [[package]]
name = "try-lock" name = "try-lock"
version = "0.2.3" version = "0.2.3"

View file

@ -9,9 +9,9 @@ description = "A MangaDex@Home implementation in Rust."
repository = "https://github.com/edward-shen/mangadex-home-rs" repository = "https://github.com/edward-shen/mangadex-home-rs"
[profile.release] [profile.release]
lto = true # lto = true
codegen-units = 1 # codegen-units = 1
# debug = 1 debug = 1
[dependencies] [dependencies]
actix-web = { version = "4.0.0-beta.4", features = [ "rustls" ] } actix-web = { version = "4.0.0-beta.4", features = [ "rustls" ] }
@ -20,6 +20,7 @@ async-trait = "0.1"
base64 = "0.13" base64 = "0.13"
bincode = "1" bincode = "1"
bytes = "1" bytes = "1"
chacha20 = "0.7"
chrono = { version = "0.4", features = [ "serde" ] } chrono = { version = "0.4", features = [ "serde" ] }
clap = { version = "3.0.0-beta.2", features = [ "wrap_help" ] } clap = { version = "3.0.0-beta.2", features = [ "wrap_help" ] }
ctrlc = "3" ctrlc = "3"
@ -38,13 +39,15 @@ serde = "1"
serde_json = "1" serde_json = "1"
serde_repr = "0.1" serde_repr = "0.1"
serde_yaml = "0.8" serde_yaml = "0.8"
simple_logger = "1"
sodiumoxide = "0.2" sodiumoxide = "0.2"
sqlx = { version = "0.5", features = [ "runtime-actix-rustls", "sqlite", "time", "chrono", "macros", "offline" ] } sqlx = { version = "0.5", features = [ "runtime-actix-rustls", "sqlite", "time", "chrono", "macros", "offline" ] }
thiserror = "1" thiserror = "1"
tokio = { version = "1", features = [ "full", "parking_lot" ] } tokio = { version = "1", features = [ "full", "parking_lot" ] }
tokio-stream = { version = "0.1", features = [ "sync" ] } tokio-stream = { version = "0.1", features = [ "sync" ] }
tokio-util = { version = "0.6", features = [ "codec" ] } tokio-util = { version = "0.6", features = [ "codec" ] }
tracing = "0.1"
tracing-futures = "0.2"
tracing-subscriber = { version = "0.2", features = [ "parking_lot" ] }
url = { version = "2", features = [ "serde" ] } url = { version = "2", features = [ "serde" ] }
[build-dependencies] [build-dependencies]

View file

@ -54,3 +54,38 @@ server_settings:
# the backend will infer it from where it was sent from, which may fail in the # the backend will infer it from where it was sent from, which may fail in the
# presence of multiple IPs. # presence of multiple IPs.
# external_ip: ~ # external_ip: ~
# These settings are unique to the Rust client, and may be ignored or behave
# differently from the official client.
extended_options:
# Which cache type to use. By default, this is `on_disk`, but one can select
# `lfu` or `lru` to use a LFU or LRU in addition to the on-disk cache to
# improve lookup times. Generally speaking, using one is almost always better,
# but by how much depends on how much memory you let the node use, how large
# is your node, and which caching implementation you use.
# cache_type: on_disk
# The amount of memory the client should use when using an in-memory cache.
# This does nothing if only the on-disk cache is used.
# memory_quota: 0
# Whether or not to expose a prometheus endpoint at /metrics. This is a
# completely open endpoint, so best practice is to make sure this is only
# readable from the internal network.
# enable_metrics: false
# If you'd like to specify a different path location for the cache, you can do
# so here.
# cache_path: "./cache"
# What logging level to use. Valid options are "error", "warn", "info",
# "debug", "trace", and "off", which disables logging.
# logging_level: info
# Warning: Experimental. Will cause problems.
#
# Enables disk encryption where the key is stored in memory. In other words,
# when the MD@H program is stopped, all cached files are irrecoverable.
# Practically speaking, this isn't all too useful (and definitely hurts
# performance), but for peace of mind, this may be useful.
# ephemeral_disk_encryption: false

14
src/cache/disk.rs vendored
View file

@ -10,7 +10,7 @@ use std::sync::Arc;
use async_trait::async_trait; use async_trait::async_trait;
use futures::StreamExt; use futures::StreamExt;
use log::{debug, error, warn, LevelFilter}; use log::LevelFilter;
use md5::digest::generic_array::GenericArray; use md5::digest::generic_array::GenericArray;
use md5::{Digest, Md5}; use md5::{Digest, Md5};
use sqlx::sqlite::SqliteConnectOptions; use sqlx::sqlite::SqliteConnectOptions;
@ -18,11 +18,13 @@ use sqlx::{ConnectOptions, Sqlite, SqlitePool, Transaction};
use tokio::fs::remove_file; use tokio::fs::remove_file;
use tokio::sync::mpsc::{channel, Receiver, Sender}; use tokio::sync::mpsc::{channel, Receiver, Sender};
use tokio_stream::wrappers::ReceiverStream; use tokio_stream::wrappers::ReceiverStream;
use tracing::{debug, error, instrument, warn};
use crate::units::Bytes; use crate::units::Bytes;
use super::{Cache, CacheError, CacheKey, CacheStream, CallbackCache, ImageMetadata}; use super::{Cache, CacheError, CacheKey, CacheStream, CallbackCache, ImageMetadata};
#[derive(Debug)]
pub struct DiskCache { pub struct DiskCache {
disk_path: PathBuf, disk_path: PathBuf,
disk_cur_size: AtomicU64, disk_cur_size: AtomicU64,
@ -214,6 +216,7 @@ async fn db_listener(
} }
} }
#[instrument(level = "debug", skip(transaction))]
async fn handle_db_get(entry: &Path, transaction: &mut Transaction<'_, Sqlite>) { async fn handle_db_get(entry: &Path, transaction: &mut Transaction<'_, Sqlite>) {
let hash = if let Ok(hash) = Md5Hash::try_from(entry) { let hash = if let Ok(hash) = Md5Hash::try_from(entry) {
hash hash
@ -241,6 +244,7 @@ async fn handle_db_get(entry: &Path, transaction: &mut Transaction<'_, Sqlite>)
} }
} }
#[instrument(level = "debug", skip(transaction, cache))]
async fn handle_db_put( async fn handle_db_put(
entry: &Path, entry: &Path,
size: u64, size: u64,
@ -263,7 +267,7 @@ async fn handle_db_put(
.await; .await;
if let Err(e) = query { if let Err(e) = query {
warn!("Failed to add {:?} to db: {}", key, e); warn!("Failed to add to db: {}", e);
} }
cache.disk_cur_size.fetch_add(size, Ordering::Release); cache.disk_cur_size.fetch_add(size, Ordering::Release);
@ -324,15 +328,15 @@ impl Cache for DiskCache {
) -> Option<Result<(CacheStream, ImageMetadata), CacheError>> { ) -> Option<Result<(CacheStream, ImageMetadata), CacheError>> {
let channel = self.db_update_channel_sender.clone(); let channel = self.db_update_channel_sender.clone();
// TODO: Check legacy path as well
let path = Arc::new(self.disk_path.clone().join(PathBuf::from(key))); let path = Arc::new(self.disk_path.clone().join(PathBuf::from(key)));
let path_0 = Arc::clone(&path); let path_0 = Arc::clone(&path);
tokio::spawn(async move { channel.send(DbMessage::Get(path_0)).await }); tokio::spawn(async move { channel.send(DbMessage::Get(path_0)).await });
super::fs::read_file_from_path(&path).await.map(|res| { super::fs::read_file_from_path(&path).await.map(|res| {
let (inner, maybe_header, metadata) = res?; res.map(|(stream, _, metadata)| (stream, metadata))
CacheStream::new(inner, maybe_header)
.map(|stream| (stream, metadata))
.map_err(|_| CacheError::DecryptionFailure) .map_err(|_| CacheError::DecryptionFailure)
}) })
} }

290
src/cache/fs.rs vendored
View file

@ -22,26 +22,24 @@ use std::pin::Pin;
use std::task::{Context, Poll}; use std::task::{Context, Poll};
use actix_web::error::PayloadError; use actix_web::error::PayloadError;
use async_trait::async_trait;
use bytes::Bytes; use bytes::Bytes;
use chacha20::cipher::{NewCipher, StreamCipher};
use chacha20::{Key, XChaCha20, XNonce};
use futures::Future; use futures::Future;
use log::{debug, warn}; use serde::Deserialize;
use serde::{Deserialize, Serialize}; use sodiumoxide::crypto::stream::xchacha20::{gen_nonce, NONCEBYTES};
use sodiumoxide::crypto::secretstream::{
Header, Pull, Push, Stream as SecretStream, Tag, HEADERBYTES,
};
use tokio::fs::{create_dir_all, remove_file, File}; use tokio::fs::{create_dir_all, remove_file, File};
use tokio::io::{AsyncRead, AsyncReadExt, AsyncSeekExt, AsyncWrite, AsyncWriteExt, ReadBuf}; use tokio::io::{
AsyncBufRead, AsyncRead, AsyncReadExt, AsyncSeekExt, AsyncWrite, AsyncWriteExt, BufReader,
ReadBuf,
};
use tokio::sync::mpsc::Sender; use tokio::sync::mpsc::Sender;
use tokio_util::codec::{BytesCodec, FramedRead}; use tokio_util::codec::{BytesCodec, FramedRead};
use tracing::{debug, instrument, warn};
use super::compat::LegacyImageMetadata; use super::compat::LegacyImageMetadata;
use super::{CacheKey, ImageMetadata, InnerStream, ENCRYPTION_KEY}; use super::{CacheKey, CacheStream, ImageMetadata, ENCRYPTION_KEY};
#[derive(Serialize, Deserialize)]
pub enum OnDiskMetadata {
Encrypted(Header, ImageMetadata),
Plaintext(ImageMetadata),
}
/// Attempts to lookup the file on disk, returning a byte stream if it exists. /// Attempts to lookup the file on disk, returning a byte stream if it exists.
/// Note that this could return two types of streams, depending on if the file /// Note that this could return two types of streams, depending on if the file
@ -49,13 +47,14 @@ pub enum OnDiskMetadata {
#[inline] #[inline]
pub(super) async fn read_file_from_path( pub(super) async fn read_file_from_path(
path: &Path, path: &Path,
) -> Option<Result<(InnerStream, Option<Header>, ImageMetadata), std::io::Error>> { ) -> Option<Result<(CacheStream, Option<XNonce>, ImageMetadata), std::io::Error>> {
read_file(std::fs::File::open(path).ok()?).await read_file(std::fs::File::open(path).ok()?).await
} }
#[instrument(level = "debug")]
async fn read_file( async fn read_file(
file: std::fs::File, file: std::fs::File,
) -> Option<Result<(InnerStream, Option<Header>, ImageMetadata), std::io::Error>> { ) -> Option<Result<(CacheStream, Option<XNonce>, ImageMetadata), std::io::Error>> {
let mut file_0 = file.try_clone().unwrap(); let mut file_0 = file.try_clone().unwrap();
let file_1 = file.try_clone().unwrap(); let file_1 = file.try_clone().unwrap();
@ -73,7 +72,7 @@ async fn read_file(
let parsed_metadata; let parsed_metadata;
let mut maybe_header = None; let mut maybe_header = None;
let mut reader: Option<Pin<Box<dyn AsyncRead + Send>>> = None; let mut reader: Option<Pin<Box<dyn MetadataFetch + Send>>> = None;
if let Ok(metadata) = maybe_metadata { if let Ok(metadata) = maybe_metadata {
// image is decrypted // image is decrypted
if ENCRYPTION_KEY.get().is_some() { if ENCRYPTION_KEY.get().is_some() {
@ -83,57 +82,55 @@ async fn read_file(
return None; return None;
} }
reader = Some(Box::pin(File::from_std(file_1))); reader = Some(Box::pin(BufReader::new(File::from_std(file_1))));
parsed_metadata = Some(metadata); parsed_metadata = Some(metadata);
debug!("Found not encrypted file"); debug!("Found not encrypted file");
} else { } else {
debug!("metadata read failed, trying to see if it's encrypted");
let mut file = File::from_std(file_1); let mut file = File::from_std(file_1);
file.seek(SeekFrom::Start(0)).await.ok()?; file.seek(SeekFrom::Start(0)).await.ok()?;
let file_0 = file.try_clone().await.unwrap();
// image is encrypted or corrupt // image is encrypted or corrupt
// If the encryption key was set, use the encrypted disk reader instead; // If the encryption key was set, use the encrypted disk reader instead;
// else, just directly read from file. // else, just directly read from file.
if let Some(key) = ENCRYPTION_KEY.get() { if let Some(key) = ENCRYPTION_KEY.get() {
let mut header_bytes = [0; HEADERBYTES]; let mut nonce_bytes = [0; NONCEBYTES];
if let Err(e) = file.read_exact(&mut header_bytes).await { if let Err(e) = file.read_exact(&mut nonce_bytes).await {
warn!("Found file but failed reading header: {}", e); warn!("Found file but failed reading header: {}", e);
return None; return None;
} }
let file_header = if let Some(header) = Header::from_slice(&header_bytes) { debug!("header bytes: {:x?}", nonce_bytes);
header
} else {
warn!("Found file, but encrypted header was invalid. Assuming corrupted!");
return None;
};
let secret_stream = if let Ok(stream) = SecretStream::init_pull(&file_header, key) { maybe_header = Some(*XNonce::from_slice(&nonce_bytes));
stream reader = Some(Box::pin(BufReader::new(EncryptedDiskReader::new(
} else { file,
warn!("Failed to init secret stream with key and header. Assuming corrupted!"); XNonce::from_slice(XNonce::from_slice(&nonce_bytes)),
return None; key,
}; ))));
maybe_header = Some(file_header);
reader = Some(Box::pin(EncryptedDiskReader::new(file, secret_stream)));
} }
let mut deserializer = serde_json::Deserializer::from_reader(file_0.into_std().await); parsed_metadata = if let Some(reader) = reader.as_mut() {
parsed_metadata = ImageMetadata::deserialize(&mut deserializer).ok(); if let Ok(metadata) = reader.as_mut().metadata().await {
debug!("Successfully parsed encrypted metadata");
if parsed_metadata.is_some() { Some(metadata)
debug!("Found encrypted file"); } else {
debug!("Failed to parse encrypted metadata");
None
} }
} else {
debug!("Failed to read encrypted data");
None
};
} }
// parsed_metadata is either set or unset here. If it's set then we // parsed_metadata is either set or unset here. If it's set then we
// successfully decoded the data; otherwise the file is garbage. // successfully decoded the data; otherwise the file is garbage.
if let Some(reader) = reader { if let Some(reader) = reader {
let stream = InnerStream::Completed(FramedRead::new(reader, BytesCodec::new())); let stream =
CacheStream::Completed(FramedRead::new(reader as Pin<Box<_>>, BytesCodec::new()));
parsed_metadata.map(|metadata| Ok((stream, maybe_header, metadata))) parsed_metadata.map(|metadata| Ok((stream, maybe_header, metadata)))
} else { } else {
debug!("Reader was invalid, file is corrupt"); debug!("Reader was invalid, file is corrupt");
@ -143,63 +140,82 @@ async fn read_file(
struct EncryptedDiskReader { struct EncryptedDiskReader {
file: Pin<Box<File>>, file: Pin<Box<File>>,
stream: SecretStream<Pull>, keystream: XChaCha20,
buf: Vec<u8>,
} }
impl EncryptedDiskReader { impl EncryptedDiskReader {
fn new(file: File, stream: SecretStream<Pull>) -> Self { fn new(file: File, nonce: &XNonce, key: &Key) -> Self {
Self { Self {
file: Box::pin(file), file: Box::pin(file),
stream, keystream: XChaCha20::new(key, nonce),
buf: vec![],
} }
} }
} }
#[async_trait]
pub trait MetadataFetch: AsyncBufRead {
async fn metadata(mut self: Pin<&mut Self>) -> Result<ImageMetadata, ()>;
}
#[async_trait]
impl<R: AsyncBufRead + Send> MetadataFetch for R {
#[inline]
async fn metadata(mut self: Pin<&mut Self>) -> Result<ImageMetadata, ()> {
MetadataFuture(&mut self).await
}
}
impl AsyncRead for EncryptedDiskReader { impl AsyncRead for EncryptedDiskReader {
fn poll_read( fn poll_read(
mut self: Pin<&mut Self>, mut self: Pin<&mut Self>,
cx: &mut Context<'_>, cx: &mut Context<'_>,
buf: &mut ReadBuf<'_>, buf: &mut ReadBuf<'_>,
) -> Poll<std::io::Result<()>> { ) -> Poll<std::io::Result<()>> {
let cursor_start = buf.filled().len(); let previously_read = buf.filled().len();
let res = self.as_mut().file.as_mut().poll_read(cx, buf); let res = self.as_mut().file.as_mut().poll_read(cx, buf);
if res.is_pending() { let bytes_modified = buf.filled().len() - previously_read;
return Poll::Pending; self.keystream.apply_keystream(
} &mut buf.filled_mut()[previously_read..previously_read + bytes_modified],
);
let cursor_new = buf.filled().len();
// pull_to_vec internally calls vec.clear() and vec.reserve(). Generally
// speaking we should be reading about the same amount of data each time
// so we shouldn't experience too much of a slow down w.r.t resizing the
// buffer...
let new_self = Pin::into_inner(self);
new_self
.stream
.pull_to_vec(
&buf.filled()[cursor_start..cursor_new],
None,
&mut new_self.buf,
)
.unwrap();
// data is strictly smaller than the encrypted stream, since you need to
// encode tags as well, so this is always safe.
// rewrite encrypted data into decrypted data
let buffer = buf.filled_mut();
for (old, new) in buffer[cursor_start..].iter_mut().zip(&new_self.buf) {
*old = *new;
}
buf.set_filled(cursor_start + new_self.buf.len());
res res
} }
} }
struct MetadataFuture<'a, R>(&'a mut Pin<&'a mut R>);
impl<'a, R: AsyncBufRead> Future for MetadataFuture<'a, R> {
type Output = Result<ImageMetadata, ()>;
fn poll(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Self::Output> {
let mut filled = 0;
loop {
let buf = match self.0.as_mut().poll_fill_buf(cx) {
Poll::Ready(Ok(buffer)) => buffer,
Poll::Ready(Err(_)) => return Poll::Ready(Err(())),
Poll::Pending => return Poll::Pending,
};
if filled == buf.len() {
return Poll::Ready(Err(()));
}
filled = buf.len();
let mut reader = serde_json::Deserializer::from_slice(buf).into_iter();
let (res, bytes_consumed) = match reader.next() {
Some(Ok(metadata)) => (Poll::Ready(Ok(metadata)), reader.byte_offset()),
Some(Err(e)) if e.is_eof() => continue,
Some(Err(_)) | None => return Poll::Ready(Err(())),
};
// This needs to be outside the loop because we need to drop the
// reader ref, since that depends on a mut self.
self.0.as_mut().consume(bytes_consumed);
return res;
}
}
}
/// Writes the metadata and input stream (in that order) to a file, returning a /// Writes the metadata and input stream (in that order) to a file, returning a
/// stream that reads from that file. Accepts a db callback function that is /// stream that reads from that file. Accepts a db callback function that is
/// provided the number of bytes written, and an optional on-complete callback /// provided the number of bytes written, and an optional on-complete callback
@ -216,36 +232,31 @@ where
Fut: 'static + Send + Sync + Future<Output = ()>, Fut: 'static + Send + Sync + Future<Output = ()>,
DbCallback: 'static + Send + Sync + FnOnce(u64) -> Fut, DbCallback: 'static + Send + Sync + FnOnce(u64) -> Fut,
{ {
let file = { let mut file = {
let parent = path.parent().expect("The path to have a parent"); let parent = path.parent().expect("The path to have a parent");
create_dir_all(parent).await?; create_dir_all(parent).await?;
let file = File::create(path).await?; // we need to make sure the file exists and is truncated. let file = File::create(path).await?; // we need to make sure the file exists and is truncated.
file file
}; };
let mut writer: Pin<Box<dyn AsyncWrite + Send>> = if let Some(key) = ENCRYPTION_KEY.get() {
let nonce = gen_nonce();
file.write_all(nonce.as_ref()).await?;
Box::pin(EncryptedDiskWriter::new(
file,
XNonce::from_slice(nonce.as_ref()),
key,
))
} else {
Box::pin(file)
};
let metadata_string = serde_json::to_string(&metadata).expect("serialization to work"); let metadata_string = serde_json::to_string(&metadata).expect("serialization to work");
let metadata_size = metadata_string.len(); let metadata_size = metadata_string.len();
let (mut writer, maybe_header): (Pin<Box<dyn AsyncWrite + Send>>, _) = let mut error = writer.write_all(metadata_string.as_bytes()).await.err();
if let Some((enc, header)) = ENCRYPTION_KEY
.get()
.map(|key| SecretStream::init_push(key).expect("Failed to init enc stream"))
{
(Box::pin(EncryptedDiskWriter::new(file, enc)), Some(header))
} else {
(Box::pin(file), None)
};
let mut error = if let Some(header) = maybe_header {
writer.write_all(header.as_ref()).await.err()
} else {
None
};
if error.is_none() {
error = writer.write_all(metadata_string.as_bytes()).await.err();
}
if error.is_none() { if error.is_none() {
debug!("decrypted write {:x?}", &bytes[..40]);
error = writer.write_all(&bytes).await.err(); error = writer.write_all(&bytes).await.err();
} }
@ -277,87 +288,76 @@ where
struct EncryptedDiskWriter { struct EncryptedDiskWriter {
file: Pin<Box<File>>, file: Pin<Box<File>>,
stream: Option<SecretStream<Push>>, keystream: XChaCha20,
encryption_buffer: Vec<u8>, buffer: Vec<u8>,
write_buffer: Vec<u8>,
} }
impl EncryptedDiskWriter { impl EncryptedDiskWriter {
fn new(file: File, stream: SecretStream<Push>) -> Self { fn new(file: File, nonce: &XNonce, key: &Key) -> Self {
Self { Self {
file: Box::pin(file), file: Box::pin(file),
stream: Some(stream), keystream: XChaCha20::new(key, nonce),
encryption_buffer: vec![], buffer: vec![],
write_buffer: vec![],
} }
} }
} }
impl AsyncWrite for EncryptedDiskWriter { impl AsyncWrite for EncryptedDiskWriter {
#[inline]
fn poll_write( fn poll_write(
self: Pin<&mut Self>, self: Pin<&mut Self>,
cx: &mut Context<'_>, cx: &mut Context<'_>,
buf: &[u8], buf: &[u8],
) -> Poll<Result<usize, std::io::Error>> { ) -> Poll<Result<usize, std::io::Error>> {
let new_self = Pin::into_inner(self); let pinned = Pin::into_inner(self);
{
let encryption_buffer = &mut new_self.encryption_buffer;
if let Some(stream) = new_self.stream.as_mut() {
stream
.push_to_vec(buf, None, Tag::Message, encryption_buffer)
.expect("Failed to write encrypted data to buffer");
}
}
new_self.write_buffer.extend(&new_self.encryption_buffer); let old_buffer_size = pinned.buffer.len();
pinned.buffer.extend_from_slice(buf);
match new_self pinned
.file .keystream
.as_mut() .apply_keystream(&mut pinned.buffer[old_buffer_size..]);
.poll_write(cx, &new_self.write_buffer) match pinned.file.as_mut().poll_write(cx, &pinned.buffer) {
{
Poll::Ready(Ok(n)) => { Poll::Ready(Ok(n)) => {
new_self.write_buffer.drain(..n); pinned.buffer.drain(..n);
// We buffered all the bytes that were provided to use.
Poll::Ready(Ok(buf.len())) Poll::Ready(Ok(buf.len()))
} }
poll => poll, Poll::Ready(Err(e)) => Poll::Ready(Err(e)),
Poll::Pending => Poll::Pending,
} }
} }
#[inline]
fn poll_flush( fn poll_flush(
mut self: Pin<&mut Self>, mut self: Pin<&mut Self>,
cx: &mut Context<'_>, cx: &mut Context<'_>,
) -> Poll<Result<(), std::io::Error>> { ) -> Poll<Result<(), std::io::Error>> {
if self.as_ref().write_buffer.is_empty() { if self.buffer.is_empty() {
self.file.as_mut().poll_flush(cx) self.as_mut().file.as_mut().poll_flush(cx)
} else { } else {
let new_self = Pin::into_inner(self); let pinned = Pin::into_inner(self);
let buffer = new_self.write_buffer.as_ref(); while !pinned.buffer.is_empty() {
match new_self.file.as_mut().poll_write(cx, buffer) { match pinned.file.as_mut().poll_write(cx, &pinned.buffer) {
Poll::Ready(res) => { Poll::Ready(Ok(n)) => {
let n = res?; pinned.buffer.drain(..n);
new_self.write_buffer.drain(..n);
// We're immediately ready to do some more flushing!
cx.waker().wake_by_ref();
// Return pending here because we still need to flush the
// file
Poll::Pending
}
Poll::Pending => Poll::Pending,
} }
Poll::Ready(Err(e)) => return Poll::Ready(Err(e)),
Poll::Pending => return Poll::Pending,
} }
} }
Poll::Ready(Ok(()))
}
}
#[inline]
fn poll_shutdown( fn poll_shutdown(
mut self: Pin<&mut Self>, mut self: Pin<&mut Self>,
cx: &mut Context<'_>, cx: &mut Context<'_>,
) -> Poll<Result<(), std::io::Error>> { ) -> Poll<Result<(), std::io::Error>> {
self.as_mut() match self.as_mut().poll_flush(cx) {
.stream Poll::Ready(Ok(())) => self.as_mut().file.as_mut().poll_shutdown(cx),
.take() poll => poll,
.map(|stream| stream.finalize(None)); }
self.file.as_mut().poll_shutdown(cx)
} }
} }

12
src/cache/mem.rs vendored
View file

@ -1,9 +1,7 @@
use std::sync::atomic::{AtomicU64, Ordering}; use std::sync::atomic::{AtomicU64, Ordering};
use std::sync::Arc; use std::sync::Arc;
use super::{ use super::{Cache, CacheKey, CacheStream, CallbackCache, ImageMetadata, MemStream};
Cache, CacheError, CacheKey, CacheStream, CallbackCache, ImageMetadata, InnerStream, MemStream,
};
use async_trait::async_trait; use async_trait::async_trait;
use bytes::Bytes; use bytes::Bytes;
use futures::FutureExt; use futures::FutureExt;
@ -146,13 +144,9 @@ where
) -> Option<Result<(CacheStream, ImageMetadata), super::CacheError>> { ) -> Option<Result<(CacheStream, ImageMetadata), super::CacheError>> {
match self.mem_cache.lock().now_or_never() { match self.mem_cache.lock().now_or_never() {
Some(mut mem_cache) => match mem_cache.get(key).map(|(bytes, metadata, _)| { Some(mut mem_cache) => match mem_cache.get(key).map(|(bytes, metadata, _)| {
Ok((InnerStream::Memory(MemStream(bytes.clone())), *metadata)) Ok((CacheStream::Memory(MemStream(bytes.clone())), *metadata))
}) { }) {
Some(v) => Some(v.and_then(|(inner, metadata)| { Some(v) => Some(v),
CacheStream::new(inner, None)
.map(|v| (v, metadata))
.map_err(|_| CacheError::DecryptionFailure)
})),
None => self.inner.get(key).await, None => self.inner.get(key).await,
}, },
None => self.inner.get(key).await, None => self.inner.get(key).await,

56
src/cache/mod.rs vendored
View file

@ -8,14 +8,13 @@ use std::task::{Context, Poll};
use actix_web::http::HeaderValue; use actix_web::http::HeaderValue;
use async_trait::async_trait; use async_trait::async_trait;
use bytes::{Bytes, BytesMut}; use bytes::{Bytes, BytesMut};
use chacha20::Key;
use chrono::{DateTime, FixedOffset}; use chrono::{DateTime, FixedOffset};
use futures::{Stream, StreamExt}; use futures::{Stream, StreamExt};
use once_cell::sync::OnceCell; use once_cell::sync::OnceCell;
use serde::{Deserialize, Serialize}; use serde::{Deserialize, Serialize};
use serde_repr::{Deserialize_repr, Serialize_repr}; use serde_repr::{Deserialize_repr, Serialize_repr};
use sodiumoxide::crypto::secretstream::{Header, Key, Pull, Stream as SecretStream};
use thiserror::Error; use thiserror::Error;
use tokio::io::AsyncRead;
use tokio::sync::mpsc::Sender; use tokio::sync::mpsc::Sender;
use tokio_util::codec::{BytesCodec, FramedRead}; use tokio_util::codec::{BytesCodec, FramedRead};
@ -24,6 +23,7 @@ pub use fs::UpstreamError;
pub use mem::MemoryCache; pub use mem::MemoryCache;
use self::compat::LegacyImageMetadata; use self::compat::LegacyImageMetadata;
use self::fs::MetadataFetch;
pub static ENCRYPTION_KEY: OnceCell<Key> = OnceCell::new(); pub static ENCRYPTION_KEY: OnceCell<Key> = OnceCell::new();
@ -231,56 +231,12 @@ impl<T: CallbackCache> CallbackCache for Arc<T> {
.await .await
} }
} }
pub enum CacheStream {
pub struct CacheStream {
inner: InnerStream,
decrypt: Option<SecretStream<Pull>>,
}
impl CacheStream {
pub(self) fn new(inner: InnerStream, header: Option<Header>) -> Result<Self, ()> {
Ok(Self {
inner,
decrypt: header
.and_then(|header| {
ENCRYPTION_KEY
.get()
.map(|key| SecretStream::init_pull(&header, key))
})
.transpose()?,
})
}
}
impl Stream for CacheStream {
type Item = CacheStreamItem;
fn poll_next(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Option<Self::Item>> {
self.inner.poll_next_unpin(cx).map(|data| {
// False positive (`data`): https://link.eddie.sh/r1fXX
#[allow(clippy::option_if_let_else)]
if let Some(keystream) = self.decrypt.as_mut() {
data.map(|bytes_res| {
bytes_res.and_then(|bytes| {
keystream
.pull(&bytes, None)
.map(|(data, _tag)| Bytes::from(data))
.map_err(|_| UpstreamError)
})
})
} else {
data
}
})
}
}
pub(self) enum InnerStream {
Memory(MemStream), Memory(MemStream),
Completed(FramedRead<Pin<Box<dyn AsyncRead + Send>>, BytesCodec>), Completed(FramedRead<Pin<Box<dyn MetadataFetch + Send>>, BytesCodec>),
} }
impl From<CachedImage> for InnerStream { impl From<CachedImage> for CacheStream {
fn from(image: CachedImage) -> Self { fn from(image: CachedImage) -> Self {
Self::Memory(MemStream(image.0)) Self::Memory(MemStream(image.0))
} }
@ -288,7 +244,7 @@ impl From<CachedImage> for InnerStream {
type CacheStreamItem = Result<Bytes, UpstreamError>; type CacheStreamItem = Result<Bytes, UpstreamError>;
impl Stream for InnerStream { impl Stream for CacheStream {
type Item = CacheStreamItem; type Item = CacheStreamItem;
fn poll_next(self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Option<Self::Item>> { fn poll_next(self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Option<Self::Item>> {

View file

@ -5,7 +5,6 @@ use std::time::Duration;
use actix_web::http::{HeaderMap, HeaderName, HeaderValue}; use actix_web::http::{HeaderMap, HeaderName, HeaderValue};
use actix_web::web::Data; use actix_web::web::Data;
use bytes::Bytes; use bytes::Bytes;
use log::{debug, error, warn};
use once_cell::sync::Lazy; use once_cell::sync::Lazy;
use parking_lot::RwLock; use parking_lot::RwLock;
use reqwest::header::{ use reqwest::header::{
@ -15,6 +14,7 @@ use reqwest::header::{
use reqwest::{Client, StatusCode}; use reqwest::{Client, StatusCode};
use tokio::sync::watch::{channel, Receiver}; use tokio::sync::watch::{channel, Receiver};
use tokio::sync::Notify; use tokio::sync::Notify;
use tracing::{debug, error, warn};
use crate::cache::{Cache, CacheKey, ImageMetadata}; use crate::cache::{Cache, CacheKey, ImageMetadata};

View file

@ -12,6 +12,7 @@ use clap::{crate_authors, crate_description, crate_version, Clap};
use log::LevelFilter; use log::LevelFilter;
use serde::{Deserialize, Serialize}; use serde::{Deserialize, Serialize};
use thiserror::Error; use thiserror::Error;
use tracing::level_filters::LevelFilter as TracingLevelFilter;
use url::Url; use url::Url;
use crate::units::{KilobitsPerSecond, Mebibytes, Port}; use crate::units::{KilobitsPerSecond, Mebibytes, Port};
@ -78,7 +79,7 @@ pub struct Config {
pub cache_type: CacheType, pub cache_type: CacheType,
pub cache_path: PathBuf, pub cache_path: PathBuf,
pub shutdown_timeout: NonZeroU16, pub shutdown_timeout: NonZeroU16,
pub log_level: LevelFilter, pub log_level: TracingLevelFilter,
pub client_secret: ClientSecret, pub client_secret: ClientSecret,
pub port: Port, pub port: Port,
pub bind_address: SocketAddr, pub bind_address: SocketAddr,
@ -97,15 +98,24 @@ impl Config {
let file_extended_options = file_args.extended_options.unwrap_or_default(); let file_extended_options = file_args.extended_options.unwrap_or_default();
let log_level = match (cli_args.quiet, cli_args.verbose) { let log_level = match (cli_args.quiet, cli_args.verbose) {
(n, _) if n > 2 => LevelFilter::Off, (n, _) if n > 2 => TracingLevelFilter::OFF,
(2, _) => LevelFilter::Error, (2, _) => TracingLevelFilter::ERROR,
(1, _) => LevelFilter::Warn, (1, _) => TracingLevelFilter::WARN,
// Use log level from file if no flags were provided to CLI // Use log level from file if no flags were provided to CLI
(0, 0) => file_extended_options (0, 0) => {
file_extended_options
.logging_level .logging_level
.unwrap_or(LevelFilter::Info), .map_or(TracingLevelFilter::INFO, |filter| match filter {
(_, 1) => LevelFilter::Debug, LevelFilter::Off => TracingLevelFilter::OFF,
(_, n) if n > 1 => LevelFilter::Trace, LevelFilter::Error => TracingLevelFilter::ERROR,
LevelFilter::Warn => TracingLevelFilter::WARN,
LevelFilter::Info => TracingLevelFilter::INFO,
LevelFilter::Debug => TracingLevelFilter::DEBUG,
LevelFilter::Trace => TracingLevelFilter::TRACE,
})
}
(_, 1) => TracingLevelFilter::DEBUG,
(_, n) if n > 1 => TracingLevelFilter::TRACE,
// compiler can't figure it out // compiler can't figure it out
_ => unsafe { unreachable_unchecked() }, _ => unsafe { unreachable_unchecked() },
}; };

View file

@ -14,15 +14,15 @@ use actix_web::rt::{spawn, time, System};
use actix_web::web::{self, Data}; use actix_web::web::{self, Data};
use actix_web::{App, HttpResponse, HttpServer}; use actix_web::{App, HttpResponse, HttpServer};
use cache::{Cache, DiskCache}; use cache::{Cache, DiskCache};
use chacha20::Key;
use config::Config; use config::Config;
use log::{debug, error, info, warn};
use parking_lot::RwLock; use parking_lot::RwLock;
use rustls::{NoClientAuth, ServerConfig}; use rustls::{NoClientAuth, ServerConfig};
use simple_logger::SimpleLogger; use sodiumoxide::crypto::stream::xchacha20::gen_key;
use sodiumoxide::crypto::secretstream::gen_key;
use state::{RwLockServerState, ServerState}; use state::{RwLockServerState, ServerState};
use stop::send_stop; use stop::send_stop;
use thiserror::Error; use thiserror::Error;
use tracing::{debug, error, info, warn};
use crate::cache::mem::{Lfu, Lru}; use crate::cache::mem::{Lfu, Lru};
use crate::cache::{MemoryCache, ENCRYPTION_KEY}; use crate::cache::{MemoryCache, ENCRYPTION_KEY};
@ -80,7 +80,10 @@ async fn main() -> Result<(), Box<dyn Error>> {
// Logging and warnings // Logging and warnings
// //
SimpleLogger::new().with_level(config.log_level).init()?; // SimpleLogger::new().with_level(config.log_level).init()?;
tracing_subscriber::fmt()
.with_max_level(config.log_level)
.init();
if let Err(e) = print_preamble_and_warnings(&config) { if let Err(e) = print_preamble_and_warnings(&config) {
error!("{}", e); error!("{}", e);
@ -94,7 +97,9 @@ async fn main() -> Result<(), Box<dyn Error>> {
if config.ephemeral_disk_encryption { if config.ephemeral_disk_encryption {
info!("Running with at-rest encryption!"); info!("Running with at-rest encryption!");
ENCRYPTION_KEY.set(gen_key()).unwrap(); ENCRYPTION_KEY
.set(*Key::from_slice(gen_key().as_ref()))
.unwrap();
} }
if config.enable_metrics { if config.enable_metrics {

View file

@ -1,7 +1,6 @@
use std::sync::atomic::Ordering; use std::sync::atomic::Ordering;
use std::{io::BufReader, sync::Arc}; use std::{io::BufReader, sync::Arc};
use log::{debug, error, info, warn};
use rustls::internal::pemfile::{certs, rsa_private_keys}; use rustls::internal::pemfile::{certs, rsa_private_keys};
use rustls::sign::{RSASigningKey, SigningKey}; use rustls::sign::{RSASigningKey, SigningKey};
use rustls::Certificate; use rustls::Certificate;
@ -9,6 +8,7 @@ use serde::de::{MapAccess, Visitor};
use serde::{Deserialize, Serialize}; use serde::{Deserialize, Serialize};
use serde_repr::Deserialize_repr; use serde_repr::Deserialize_repr;
use sodiumoxide::crypto::box_::PrecomputedKey; use sodiumoxide::crypto::box_::PrecomputedKey;
use tracing::{debug, error, info, warn};
use url::Url; use url::Url;
use crate::config::{ClientSecret, Config, UnstableOptions, VALIDATE_TOKENS}; use crate::config::{ClientSecret, Config, UnstableOptions, VALIDATE_TOKENS};

View file

@ -10,11 +10,11 @@ use base64::DecodeError;
use bytes::Bytes; use bytes::Bytes;
use chrono::{DateTime, Utc}; use chrono::{DateTime, Utc};
use futures::Stream; use futures::Stream;
use log::{debug, error, info, trace};
use prometheus::{Encoder, TextEncoder}; use prometheus::{Encoder, TextEncoder};
use serde::Deserialize; use serde::Deserialize;
use sodiumoxide::crypto::box_::{open_precomputed, Nonce, PrecomputedKey, NONCEBYTES}; use sodiumoxide::crypto::box_::{open_precomputed, Nonce, PrecomputedKey, NONCEBYTES};
use thiserror::Error; use thiserror::Error;
use tracing::{debug, error, info, trace};
use crate::cache::{Cache, CacheKey, ImageMetadata, UpstreamError}; use crate::cache::{Cache, CacheKey, ImageMetadata, UpstreamError};
use crate::client::{FetchResult, DEFAULT_HEADERS, HTTP_CLIENT}; use crate::client::{FetchResult, DEFAULT_HEADERS, HTTP_CLIENT};

View file

@ -4,7 +4,6 @@ use std::sync::atomic::{AtomicBool, Ordering};
use crate::config::{ClientSecret, Config, UnstableOptions, OFFLINE_MODE, VALIDATE_TOKENS}; use crate::config::{ClientSecret, Config, UnstableOptions, OFFLINE_MODE, VALIDATE_TOKENS};
use crate::ping::{Request, Response, CONTROL_CENTER_PING_URL}; use crate::ping::{Request, Response, CONTROL_CENTER_PING_URL};
use arc_swap::ArcSwap; use arc_swap::ArcSwap;
use log::{error, info, warn};
use once_cell::sync::OnceCell; use once_cell::sync::OnceCell;
use parking_lot::RwLock; use parking_lot::RwLock;
use rustls::sign::{CertifiedKey, SigningKey}; use rustls::sign::{CertifiedKey, SigningKey};
@ -12,6 +11,7 @@ use rustls::Certificate;
use rustls::{ClientHello, ResolvesServerCert}; use rustls::{ClientHello, ResolvesServerCert};
use sodiumoxide::crypto::box_::{PrecomputedKey, PRECOMPUTEDKEYBYTES}; use sodiumoxide::crypto::box_::{PrecomputedKey, PRECOMPUTEDKEYBYTES};
use thiserror::Error; use thiserror::Error;
use tracing::{error, info, warn};
use url::Url; use url::Url;
pub struct ServerState { pub struct ServerState {

View file

@ -1,6 +1,6 @@
use log::{info, warn};
use reqwest::StatusCode; use reqwest::StatusCode;
use serde::Serialize; use serde::Serialize;
use tracing::{info, warn};
use crate::config::ClientSecret; use crate::config::ClientSecret;