use std::sync::atomic::Ordering; use actix_web::error::ErrorNotFound; use actix_web::http::header::{ ACCESS_CONTROL_ALLOW_ORIGIN, ACCESS_CONTROL_EXPOSE_HEADERS, CACHE_CONTROL, CONTENT_LENGTH, CONTENT_TYPE, LAST_MODIFIED, X_CONTENT_TYPE_OPTIONS, }; use actix_web::web::Path; use actix_web::HttpResponseBuilder; use actix_web::{get, web::Data, HttpRequest, HttpResponse, Responder}; use base64::DecodeError; use bytes::Bytes; use chrono::{DateTime, Utc}; use futures::Stream; use log::{debug, error, info, trace}; use prometheus::{Encoder, TextEncoder}; use serde::Deserialize; use sodiumoxide::crypto::box_::{open_precomputed, Nonce, PrecomputedKey, NONCEBYTES}; use thiserror::Error; use crate::cache::{Cache, CacheKey, ImageMetadata, UpstreamError}; use crate::client::{FetchResult, HTTP_CLIENT}; use crate::config::{OFFLINE_MODE, VALIDATE_TOKENS}; use crate::metrics::{ CACHE_HIT_COUNTER, CACHE_MISS_COUNTER, REQUESTS_DATA_COUNTER, REQUESTS_DATA_SAVER_COUNTER, REQUESTS_OTHER_COUNTER, REQUESTS_TOTAL_COUNTER, }; use crate::state::RwLockServerState; const BASE64_CONFIG: base64::Config = base64::Config::new(base64::CharacterSet::UrlSafe, false); pub enum ServerResponse { TokenValidationError(TokenValidationError), HttpResponse(HttpResponse), } impl Responder for ServerResponse { #[inline] fn respond_to(self, req: &HttpRequest) -> HttpResponse { match self { Self::TokenValidationError(e) => e.respond_to(req), Self::HttpResponse(resp) => { REQUESTS_TOTAL_COUNTER.inc(); resp.respond_to(req) } } } } #[allow(clippy::unused_async)] #[get("/")] async fn index() -> impl Responder { HttpResponse::Ok().body(include_str!("index.html")) } #[allow(clippy::future_not_send)] #[get("/{token}/data/{chapter_hash}/{file_name}")] async fn token_data( state: Data, cache: Data, path: Path<(String, String, String)>, ) -> impl Responder { REQUESTS_DATA_COUNTER.inc(); let (token, chapter_hash, file_name) = path.into_inner(); if VALIDATE_TOKENS.load(Ordering::Acquire) { if let Err(e) = validate_token(&state.0.read().precomputed_key, token, &chapter_hash) { return ServerResponse::TokenValidationError(e); } } fetch_image(state, cache, chapter_hash, file_name, false).await } #[allow(clippy::future_not_send)] #[get("/{token}/data-saver/{chapter_hash}/{file_name}")] async fn token_data_saver( state: Data, cache: Data, path: Path<(String, String, String)>, ) -> impl Responder { REQUESTS_DATA_SAVER_COUNTER.inc(); let (token, chapter_hash, file_name) = path.into_inner(); if VALIDATE_TOKENS.load(Ordering::Acquire) { if let Err(e) = validate_token(&state.0.read().precomputed_key, token, &chapter_hash) { return ServerResponse::TokenValidationError(e); } } fetch_image(state, cache, chapter_hash, file_name, true).await } #[allow(clippy::future_not_send)] pub async fn default(state: Data, req: HttpRequest) -> impl Responder { REQUESTS_OTHER_COUNTER.inc(); let path = &format!( "{}{}", state.0.read().image_server, req.path().chars().skip(1).collect::() ); if OFFLINE_MODE.load(Ordering::Acquire) { info!("Got unknown path in offline mode, returning 404: {}", path); return ServerResponse::HttpResponse( ErrorNotFound("Path is not valid in offline mode").into(), ); } info!("Got unknown path, just proxying: {}", path); let resp = match HTTP_CLIENT.inner().get(path).send().await { Ok(resp) => resp, Err(e) => { error!("{}", e); return ServerResponse::HttpResponse(HttpResponse::BadGateway().finish()); } }; let content_type = resp.headers().get(CONTENT_TYPE); let mut resp_builder = HttpResponseBuilder::new(resp.status()); if let Some(content_type) = content_type { resp_builder.insert_header((CONTENT_TYPE, content_type)); } push_headers(&mut resp_builder); ServerResponse::HttpResponse(resp_builder.body(resp.bytes().await.unwrap_or_default())) } #[allow(clippy::future_not_send, clippy::unused_async)] #[get("/metrics")] pub async fn metrics() -> impl Responder { let metric_families = prometheus::gather(); let mut buffer = Vec::new(); TextEncoder::new() .encode(&metric_families, &mut buffer) .unwrap(); String::from_utf8(buffer).unwrap() } #[derive(Error, Debug)] pub enum TokenValidationError { #[error("Failed to decode base64 token.")] DecodeError(#[from] DecodeError), #[error("Nonce was too short.")] IncompleteNonce, #[error("Invalid nonce.")] InvalidNonce, #[error("Decryption failed")] DecryptionFailure, #[error("The token format was invalid.")] InvalidToken, #[error("The token has expired.")] TokenExpired, #[error("Invalid chapter hash.")] InvalidChapterHash, } impl Responder for TokenValidationError { #[inline] fn respond_to(self, _: &HttpRequest) -> HttpResponse { push_headers(&mut HttpResponse::Forbidden()).finish() } } fn validate_token( precomputed_key: &PrecomputedKey, token: String, chapter_hash: &str, ) -> Result<(), TokenValidationError> { #[derive(Deserialize)] struct Token<'a> { expires: DateTime, hash: &'a str, } let data = base64::decode_config(token, BASE64_CONFIG)?; if data.len() < NONCEBYTES { return Err(TokenValidationError::IncompleteNonce); } let (nonce, encrypted) = data.split_at(NONCEBYTES); let nonce = Nonce::from_slice(nonce).ok_or(TokenValidationError::InvalidNonce)?; let decrypted = open_precomputed(encrypted, &nonce, precomputed_key) .map_err(|_| TokenValidationError::DecryptionFailure)?; let parsed_token: Token = serde_json::from_slice(&decrypted).map_err(|_| TokenValidationError::InvalidToken)?; if parsed_token.expires < Utc::now() { return Err(TokenValidationError::TokenExpired); } if parsed_token.hash != chapter_hash { return Err(TokenValidationError::InvalidChapterHash); } debug!("Token validated!"); Ok(()) } #[inline] pub fn push_headers(builder: &mut HttpResponseBuilder) -> &mut HttpResponseBuilder { builder .insert_header((X_CONTENT_TYPE_OPTIONS, "nosniff")) .insert_header((ACCESS_CONTROL_ALLOW_ORIGIN, "https://mangadex.org")) .insert_header((ACCESS_CONTROL_EXPOSE_HEADERS, "*")) .insert_header((CACHE_CONTROL, "public, max-age=1209600")) .insert_header(("Timing-Allow-Origin", "https://mangadex.org")); builder } #[allow(clippy::future_not_send)] async fn fetch_image( state: Data, cache: Data, chapter_hash: String, file_name: String, is_data_saver: bool, ) -> ServerResponse { let key = CacheKey(chapter_hash, file_name, is_data_saver); match cache.get(&key).await { Some(Ok((image, metadata))) => { CACHE_HIT_COUNTER.inc(); return construct_response(image, &metadata); } Some(Err(_)) => { return ServerResponse::HttpResponse(HttpResponse::BadGateway().finish()); } None => (), } CACHE_MISS_COUNTER.inc(); // If in offline mode, return early since there's nothing else we can do if OFFLINE_MODE.load(Ordering::Acquire) { return ServerResponse::HttpResponse( ErrorNotFound("Offline mode enabled and image not in cache").into(), ); } let url = if is_data_saver { format!( "{}/data-saver/{}/{}", state.0.read().image_server, &key.0, &key.1, ) } else { format!("{}/data/{}/{}", state.0.read().image_server, &key.0, &key.1) }; match HTTP_CLIENT.fetch_and_cache(url, key, cache).await { FetchResult::ServiceUnavailable => { ServerResponse::HttpResponse(HttpResponse::ServiceUnavailable().finish()) } FetchResult::InternalServerError => { ServerResponse::HttpResponse(HttpResponse::InternalServerError().finish()) } FetchResult::Data(status, headers, data) => { let mut resp = HttpResponseBuilder::new(status); let mut resp = resp.body(data); *resp.headers_mut() = headers; ServerResponse::HttpResponse(resp) } FetchResult::Processing => panic!("Race condition found with fetch result"), } } pub fn construct_response( data: impl Stream> + Unpin + 'static, metadata: &ImageMetadata, ) -> ServerResponse { trace!("Constructing response"); let mut resp = HttpResponse::Ok(); if let Some(content_type) = metadata.content_type { resp.append_header((CONTENT_TYPE, content_type.as_ref())); } if let Some(content_length) = metadata.content_length { resp.append_header((CONTENT_LENGTH, content_length)); } if let Some(last_modified) = metadata.last_modified { resp.append_header((LAST_MODIFIED, last_modified.to_rfc2822())); } ServerResponse::HttpResponse(push_headers(&mut resp).streaming(data)) }