Compare commits
No commits in common. "8a08e8e100701f0ac41ba89f97f5c024f62e5a5b" and "364a4676268bd658a72be3650204f93ef9b3604f" have entirely different histories.
8a08e8e100
...
364a467626
6 changed files with 48 additions and 119 deletions
1
Cargo.lock
generated
1
Cargo.lock
generated
|
@ -947,7 +947,6 @@ dependencies = [
|
||||||
"bytes",
|
"bytes",
|
||||||
"chrono",
|
"chrono",
|
||||||
"headers",
|
"headers",
|
||||||
"lazy_static",
|
|
||||||
"omegaupload-common",
|
"omegaupload-common",
|
||||||
"rand",
|
"rand",
|
||||||
"rocksdb",
|
"rocksdb",
|
||||||
|
|
|
@ -7,9 +7,7 @@ use anyhow::{anyhow, bail, Context, Result};
|
||||||
use atty::Stream;
|
use atty::Stream;
|
||||||
use clap::Parser;
|
use clap::Parser;
|
||||||
use omegaupload_common::crypto::{gen_key_nonce, open_in_place, seal_in_place, Key};
|
use omegaupload_common::crypto::{gen_key_nonce, open_in_place, seal_in_place, Key};
|
||||||
use omegaupload_common::{
|
use omegaupload_common::{base64, hash, Expiration, ParsedUrl, Url, API_ENDPOINT};
|
||||||
base64, hash, Expiration, ParsedUrl, Url, API_ENDPOINT, EXPIRATION_HEADER_NAME,
|
|
||||||
};
|
|
||||||
use reqwest::blocking::Client;
|
use reqwest::blocking::Client;
|
||||||
use reqwest::header::EXPIRES;
|
use reqwest::header::EXPIRES;
|
||||||
use reqwest::StatusCode;
|
use reqwest::StatusCode;
|
||||||
|
@ -30,8 +28,6 @@ enum Action {
|
||||||
/// public access.
|
/// public access.
|
||||||
#[clap(short, long)]
|
#[clap(short, long)]
|
||||||
password: Option<SecretString>,
|
password: Option<SecretString>,
|
||||||
#[clap(short, long)]
|
|
||||||
duration: Option<Expiration>,
|
|
||||||
},
|
},
|
||||||
Download {
|
Download {
|
||||||
/// The paste to download.
|
/// The paste to download.
|
||||||
|
@ -43,22 +39,14 @@ fn main() -> Result<()> {
|
||||||
let opts = Opts::parse();
|
let opts = Opts::parse();
|
||||||
|
|
||||||
match opts.action {
|
match opts.action {
|
||||||
Action::Upload {
|
Action::Upload { url, password } => handle_upload(url, password),
|
||||||
url,
|
|
||||||
password,
|
|
||||||
duration,
|
|
||||||
} => handle_upload(url, password, duration),
|
|
||||||
Action::Download { url } => handle_download(url),
|
Action::Download { url } => handle_download(url),
|
||||||
}?;
|
}?;
|
||||||
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
fn handle_upload(
|
fn handle_upload(mut url: Url, password: Option<SecretString>) -> Result<()> {
|
||||||
mut url: Url,
|
|
||||||
password: Option<SecretString>,
|
|
||||||
duration: Option<Expiration>,
|
|
||||||
) -> Result<()> {
|
|
||||||
url.set_fragment(None);
|
url.set_fragment(None);
|
||||||
|
|
||||||
if atty::is(Stream::Stdin) {
|
if atty::is(Stream::Stdin) {
|
||||||
|
@ -88,13 +76,11 @@ fn handle_upload(
|
||||||
(container, nonce, key, pw_used)
|
(container, nonce, key, pw_used)
|
||||||
};
|
};
|
||||||
|
|
||||||
let mut res = Client::new().post(url.as_ref());
|
let res = Client::new()
|
||||||
|
.post(url.as_ref())
|
||||||
if let Some(duration) = duration {
|
.body(data)
|
||||||
res = res.header(&*EXPIRATION_HEADER_NAME, duration);
|
.send()
|
||||||
}
|
.context("Request to server failed")?;
|
||||||
|
|
||||||
let res = res.body(data).send().context("Request to server failed")?;
|
|
||||||
|
|
||||||
if res.status() != StatusCode::OK {
|
if res.status() != StatusCode::OK {
|
||||||
bail!("Upload failed. Got HTTP error {}", res.status());
|
bail!("Upload failed. Got HTTP error {}", res.status());
|
||||||
|
@ -118,8 +104,11 @@ fn handle_upload(
|
||||||
}
|
}
|
||||||
|
|
||||||
fn handle_download(mut url: ParsedUrl) -> Result<()> {
|
fn handle_download(mut url: ParsedUrl) -> Result<()> {
|
||||||
url.sanitized_url
|
url.sanitized_url.set_path(&dbg!(format!(
|
||||||
.set_path(&format!("{}{}", API_ENDPOINT, url.sanitized_url.path()));
|
"{}{}",
|
||||||
|
API_ENDPOINT,
|
||||||
|
url.sanitized_url.path()
|
||||||
|
)));
|
||||||
let res = Client::new()
|
let res = Client::new()
|
||||||
.get(url.sanitized_url)
|
.get(url.sanitized_url)
|
||||||
.send()
|
.send()
|
||||||
|
|
|
@ -224,31 +224,13 @@ impl FromStr for ParsedUrl {
|
||||||
#[derive(Serialize, Deserialize, Clone, Copy, Debug)]
|
#[derive(Serialize, Deserialize, Clone, Copy, Debug)]
|
||||||
pub enum Expiration {
|
pub enum Expiration {
|
||||||
BurnAfterReading,
|
BurnAfterReading,
|
||||||
BurnAfterReadingWithDeadline(DateTime<Utc>),
|
|
||||||
UnixTime(DateTime<Utc>),
|
UnixTime(DateTime<Utc>),
|
||||||
}
|
}
|
||||||
|
|
||||||
// This impl is used for the CLI
|
|
||||||
impl FromStr for Expiration {
|
|
||||||
type Err = String;
|
|
||||||
|
|
||||||
fn from_str(s: &str) -> Result<Self, Self::Err> {
|
|
||||||
match s {
|
|
||||||
"read" => Ok(Self::BurnAfterReading),
|
|
||||||
"5m" => Ok(Self::UnixTime(Utc::now() + Duration::minutes(5))),
|
|
||||||
"10m" => Ok(Self::UnixTime(Utc::now() + Duration::minutes(10))),
|
|
||||||
"1h" => Ok(Self::UnixTime(Utc::now() + Duration::hours(1))),
|
|
||||||
"1d" => Ok(Self::UnixTime(Utc::now() + Duration::days(1))),
|
|
||||||
// We disallow permanent pastes
|
|
||||||
_ => Err(s.to_owned()),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl Display for Expiration {
|
impl Display for Expiration {
|
||||||
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||||||
match self {
|
match self {
|
||||||
Expiration::BurnAfterReading | Expiration::BurnAfterReadingWithDeadline(_) => {
|
Expiration::BurnAfterReading => {
|
||||||
write!(f, "This item has been burned. You now have the only copy.")
|
write!(f, "This item has been burned. You now have the only copy.")
|
||||||
}
|
}
|
||||||
Expiration::UnixTime(time) => write!(
|
Expiration::UnixTime(time) => write!(
|
||||||
|
@ -274,9 +256,19 @@ impl Header for Expiration {
|
||||||
Self: Sized,
|
Self: Sized,
|
||||||
I: Iterator<Item = &'i HeaderValue>,
|
I: Iterator<Item = &'i HeaderValue>,
|
||||||
{
|
{
|
||||||
let bytes = values.next().ok_or_else(headers::Error::invalid)?;
|
match values
|
||||||
|
.next()
|
||||||
Self::try_from(bytes).map_err(|_| headers::Error::invalid())
|
.ok_or_else(headers::Error::invalid)?
|
||||||
|
.as_bytes()
|
||||||
|
{
|
||||||
|
b"read" => Ok(Self::BurnAfterReading),
|
||||||
|
b"5m" => Ok(Self::UnixTime(Utc::now() + Duration::minutes(5))),
|
||||||
|
b"10m" => Ok(Self::UnixTime(Utc::now() + Duration::minutes(10))),
|
||||||
|
b"1h" => Ok(Self::UnixTime(Utc::now() + Duration::hours(1))),
|
||||||
|
b"1d" => Ok(Self::UnixTime(Utc::now() + Duration::days(1))),
|
||||||
|
// We disallow permanent pastes
|
||||||
|
_ => Err(headers::Error::invalid()),
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn encode<E: Extend<HeaderValue>>(&self, container: &mut E) {
|
fn encode<E: Extend<HeaderValue>>(&self, container: &mut E) {
|
||||||
|
@ -290,9 +282,7 @@ impl From<&Expiration> for HeaderValue {
|
||||||
// so we don't need the extra check.
|
// so we don't need the extra check.
|
||||||
unsafe {
|
unsafe {
|
||||||
Self::from_maybe_shared_unchecked(match expiration {
|
Self::from_maybe_shared_unchecked(match expiration {
|
||||||
Expiration::BurnAfterReadingWithDeadline(_) | Expiration::BurnAfterReading => {
|
Expiration::BurnAfterReading => Bytes::from_static(b"0"),
|
||||||
Bytes::from_static(b"0")
|
|
||||||
}
|
|
||||||
Expiration::UnixTime(duration) => Bytes::from(duration.to_rfc3339()),
|
Expiration::UnixTime(duration) => Bytes::from(duration.to_rfc3339()),
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
@ -305,8 +295,6 @@ impl From<Expiration> for HeaderValue {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub struct ParseHeaderValueError;
|
|
||||||
|
|
||||||
#[cfg(feature = "wasm")]
|
#[cfg(feature = "wasm")]
|
||||||
impl TryFrom<web_sys::Headers> for Expiration {
|
impl TryFrom<web_sys::Headers> for Expiration {
|
||||||
type Error = ParseHeaderValueError;
|
type Error = ParseHeaderValueError;
|
||||||
|
@ -322,19 +310,14 @@ impl TryFrom<web_sys::Headers> for Expiration {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl TryFrom<HeaderValue> for Expiration {
|
pub struct ParseHeaderValueError;
|
||||||
type Error = ParseHeaderValueError;
|
|
||||||
|
|
||||||
fn try_from(value: HeaderValue) -> Result<Self, Self::Error> {
|
|
||||||
Self::try_from(&value)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl TryFrom<&HeaderValue> for Expiration {
|
impl TryFrom<&HeaderValue> for Expiration {
|
||||||
type Error = ParseHeaderValueError;
|
type Error = ParseHeaderValueError;
|
||||||
|
|
||||||
fn try_from(value: &HeaderValue) -> Result<Self, Self::Error> {
|
fn try_from(value: &HeaderValue) -> Result<Self, Self::Error> {
|
||||||
std::str::from_utf8(value.as_bytes())
|
value
|
||||||
|
.to_str()
|
||||||
.map_err(|_| ParseHeaderValueError)
|
.map_err(|_| ParseHeaderValueError)
|
||||||
.and_then(Self::try_from)
|
.and_then(Self::try_from)
|
||||||
}
|
}
|
||||||
|
@ -344,10 +327,6 @@ impl TryFrom<&str> for Expiration {
|
||||||
type Error = ParseHeaderValueError;
|
type Error = ParseHeaderValueError;
|
||||||
|
|
||||||
fn try_from(value: &str) -> Result<Self, Self::Error> {
|
fn try_from(value: &str) -> Result<Self, Self::Error> {
|
||||||
if value == "0" {
|
|
||||||
return Ok(Self::BurnAfterReading);
|
|
||||||
}
|
|
||||||
|
|
||||||
value
|
value
|
||||||
.parse::<DateTime<Utc>>()
|
.parse::<DateTime<Utc>>()
|
||||||
.map_err(|_| ParseHeaderValueError)
|
.map_err(|_| ParseHeaderValueError)
|
||||||
|
|
|
@ -16,7 +16,6 @@ bytes = { version = "*", features = ["serde"] }
|
||||||
chrono = { version = "0.4", features = ["serde"] }
|
chrono = { version = "0.4", features = ["serde"] }
|
||||||
# We just need to pull in whatever axum is pulling in
|
# We just need to pull in whatever axum is pulling in
|
||||||
headers = "*"
|
headers = "*"
|
||||||
lazy_static = "1"
|
|
||||||
rand = "0.8"
|
rand = "0.8"
|
||||||
rocksdb = { version = "0.17", default_features = false, features = ["zstd"] }
|
rocksdb = { version = "0.17", default_features = false, features = ["zstd"] }
|
||||||
serde = { version = "1", features = ["derive"] }
|
serde = { version = "1", features = ["derive"] }
|
||||||
|
|
|
@ -14,7 +14,6 @@ use axum::response::Html;
|
||||||
use axum::{service, AddExtensionLayer, Router};
|
use axum::{service, AddExtensionLayer, Router};
|
||||||
use chrono::Utc;
|
use chrono::Utc;
|
||||||
use headers::HeaderMap;
|
use headers::HeaderMap;
|
||||||
use lazy_static::lazy_static;
|
|
||||||
use omegaupload_common::{Expiration, API_ENDPOINT};
|
use omegaupload_common::{Expiration, API_ENDPOINT};
|
||||||
use rand::thread_rng;
|
use rand::thread_rng;
|
||||||
use rand::Rng;
|
use rand::Rng;
|
||||||
|
@ -32,10 +31,6 @@ mod short_code;
|
||||||
const BLOB_CF_NAME: &str = "blob";
|
const BLOB_CF_NAME: &str = "blob";
|
||||||
const META_CF_NAME: &str = "meta";
|
const META_CF_NAME: &str = "meta";
|
||||||
|
|
||||||
lazy_static! {
|
|
||||||
static ref MAX_PASTE_AGE: chrono::Duration = chrono::Duration::days(1);
|
|
||||||
}
|
|
||||||
|
|
||||||
#[tokio::main]
|
#[tokio::main]
|
||||||
async fn main() -> Result<()> {
|
async fn main() -> Result<()> {
|
||||||
const PASTE_DB_PATH: &str = "database";
|
const PASTE_DB_PATH: &str = "database";
|
||||||
|
@ -117,10 +112,8 @@ fn set_up_expirations(db: &Arc<DB>) {
|
||||||
|
|
||||||
let expiration_time = match expiration {
|
let expiration_time = match expiration {
|
||||||
Expiration::BurnAfterReading => {
|
Expiration::BurnAfterReading => {
|
||||||
warn!("Found unbounded burn after reading. Defaulting to max age");
|
panic!("Got burn after reading expiration time? Invariant violated");
|
||||||
Utc::now() + *MAX_PASTE_AGE
|
|
||||||
}
|
}
|
||||||
Expiration::BurnAfterReadingWithDeadline(deadline) => deadline,
|
|
||||||
Expiration::UnixTime(time) => time,
|
Expiration::UnixTime(time) => time,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -159,15 +152,6 @@ async fn upload<const N: usize>(
|
||||||
return Err(StatusCode::BAD_REQUEST);
|
return Err(StatusCode::BAD_REQUEST);
|
||||||
}
|
}
|
||||||
|
|
||||||
if let Some(header) = maybe_expires {
|
|
||||||
if let Expiration::UnixTime(time) = header.0 {
|
|
||||||
if (time - Utc::now()) > *MAX_PASTE_AGE {
|
|
||||||
warn!("{} exceeds allowed paste lifetime", time);
|
|
||||||
return Err(StatusCode::BAD_REQUEST);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// 3GB max; this is a soft-limit of RocksDb
|
// 3GB max; this is a soft-limit of RocksDb
|
||||||
if body.len() >= 3_221_225_472 {
|
if body.len() >= 3_221_225_472 {
|
||||||
return Err(StatusCode::PAYLOAD_TOO_LARGE);
|
return Err(StatusCode::PAYLOAD_TOO_LARGE);
|
||||||
|
@ -201,6 +185,10 @@ async fn upload<const N: usize>(
|
||||||
return Err(StatusCode::INTERNAL_SERVER_ERROR);
|
return Err(StatusCode::INTERNAL_SERVER_ERROR);
|
||||||
};
|
};
|
||||||
|
|
||||||
|
trace!("Serializing paste...");
|
||||||
|
|
||||||
|
trace!("Finished serializing paste.");
|
||||||
|
|
||||||
let db_ref = Arc::clone(&db);
|
let db_ref = Arc::clone(&db);
|
||||||
match task::spawn_blocking(move || {
|
match task::spawn_blocking(move || {
|
||||||
let blob_cf = db_ref.cf_handle(BLOB_CF_NAME).unwrap();
|
let blob_cf = db_ref.cf_handle(BLOB_CF_NAME).unwrap();
|
||||||
|
@ -208,11 +196,6 @@ async fn upload<const N: usize>(
|
||||||
let data = bincode::serialize(&body).expect("bincode to serialize");
|
let data = bincode::serialize(&body).expect("bincode to serialize");
|
||||||
db_ref.put_cf(blob_cf, key, data)?;
|
db_ref.put_cf(blob_cf, key, data)?;
|
||||||
let expires = maybe_expires.map(|v| v.0).unwrap_or_default();
|
let expires = maybe_expires.map(|v| v.0).unwrap_or_default();
|
||||||
let expires = if let Expiration::BurnAfterReading = expires {
|
|
||||||
Expiration::BurnAfterReadingWithDeadline(Utc::now() + *MAX_PASTE_AGE)
|
|
||||||
} else {
|
|
||||||
expires
|
|
||||||
};
|
|
||||||
let meta = bincode::serialize(&expires).expect("bincode to serialize");
|
let meta = bincode::serialize(&expires).expect("bincode to serialize");
|
||||||
if db_ref.put_cf(meta_cf, key, meta).is_err() {
|
if db_ref.put_cf(meta_cf, key, meta).is_err() {
|
||||||
// try and roll back on metadata write failure
|
// try and roll back on metadata write failure
|
||||||
|
@ -224,9 +207,7 @@ async fn upload<const N: usize>(
|
||||||
{
|
{
|
||||||
Ok(Ok(_)) => {
|
Ok(Ok(_)) => {
|
||||||
if let Some(expires) = maybe_expires {
|
if let Some(expires) = maybe_expires {
|
||||||
if let Expiration::UnixTime(expiration_time)
|
if let Expiration::UnixTime(expiration_time) = expires.0 {
|
||||||
| Expiration::BurnAfterReadingWithDeadline(expiration_time) = expires.0
|
|
||||||
{
|
|
||||||
let sleep_duration =
|
let sleep_duration =
|
||||||
(expiration_time - Utc::now()).to_std().unwrap_or_default();
|
(expiration_time - Utc::now()).to_std().unwrap_or_default();
|
||||||
|
|
||||||
|
@ -321,33 +302,16 @@ async fn paste<const N: usize>(
|
||||||
};
|
};
|
||||||
|
|
||||||
// Check if we need to burn after read
|
// Check if we need to burn after read
|
||||||
if matches!(
|
if matches!(metadata, Expiration::BurnAfterReading) {
|
||||||
metadata,
|
let join_handle = task::spawn_blocking(move || db.delete(key))
|
||||||
Expiration::BurnAfterReading | Expiration::BurnAfterReadingWithDeadline(_)
|
|
||||||
) {
|
|
||||||
let join_handle = task::spawn_blocking(move || {
|
|
||||||
let blob_cf = db.cf_handle(BLOB_CF_NAME).unwrap();
|
|
||||||
let meta_cf = db.cf_handle(META_CF_NAME).unwrap();
|
|
||||||
if let Err(e) = db.delete_cf(blob_cf, url.as_bytes()) {
|
|
||||||
warn!("{}", e);
|
|
||||||
return Err(());
|
|
||||||
}
|
|
||||||
|
|
||||||
if let Err(e) = db.delete_cf(meta_cf, url.as_bytes()) {
|
|
||||||
warn!("{}", e);
|
|
||||||
return Err(());
|
|
||||||
}
|
|
||||||
|
|
||||||
Ok(())
|
|
||||||
})
|
|
||||||
.await
|
.await
|
||||||
.map_err(|e| {
|
.map_err(|e| {
|
||||||
error!("Failed to join handle: {}", e);
|
error!("Failed to join handle: {}", e);
|
||||||
StatusCode::INTERNAL_SERVER_ERROR
|
StatusCode::INTERNAL_SERVER_ERROR
|
||||||
})?;
|
})?;
|
||||||
|
|
||||||
join_handle.map_err(|_| {
|
join_handle.map_err(|e| {
|
||||||
error!("Failed to burn paste after read");
|
error!("Failed to burn paste after read: {}", e);
|
||||||
StatusCode::INTERNAL_SERVER_ERROR
|
StatusCode::INTERNAL_SERVER_ERROR
|
||||||
})?;
|
})?;
|
||||||
}
|
}
|
||||||
|
|
|
@ -113,11 +113,10 @@ pub fn decrypt(
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
entries.sort_by(|a, b| a.name.cmp(&b.name));
|
|
||||||
Ok(DecryptedData::Archive(blob, entries))
|
Ok(DecryptedData::Archive(blob, entries))
|
||||||
} else if mime_type == "application/gzip" {
|
} else if mime_type == "application/gzip" {
|
||||||
Ok(DecryptedData::Archive(blob, vec![]))
|
let entries = vec![];
|
||||||
|
Ok(DecryptedData::Archive(blob, entries))
|
||||||
} else {
|
} else {
|
||||||
Ok(DecryptedData::Blob(blob))
|
Ok(DecryptedData::Blob(blob))
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in a new issue