uhhhh some stuff rpolly?
This commit is contained in:
parent
300c0385e8
commit
33ad2f14a7
4 changed files with 147 additions and 146 deletions
40
src/cache.rs
Normal file
40
src/cache.rs
Normal file
|
@ -0,0 +1,40 @@
|
|||
use crate::state::AppState;
|
||||
|
||||
use super::error;
|
||||
|
||||
use serde::Serialize;
|
||||
use tokio::io::AsyncReadExt;
|
||||
|
||||
use std::io;
|
||||
|
||||
use std::collections::HashMap;
|
||||
|
||||
pub async fn write_to_cache<T, Y>(records: &HashMap<T, Y>) -> io::Result<()>
|
||||
where
|
||||
T: Serialize,
|
||||
Y: Serialize,
|
||||
{
|
||||
let mut records_cache = tokio::fs::File::create(".cache/data").await.unwrap();
|
||||
|
||||
let mut buf: Vec<u8> = Vec::with_capacity(200);
|
||||
bincode::serialize_into(&mut buf, records).map_err(|err| error::io_other(&err.to_string()))?;
|
||||
|
||||
let bytes_written = tokio::io::copy(&mut buf.as_slice(), &mut records_cache).await?;
|
||||
|
||||
tracing::debug!("state cache size: {}", bytes_written);
|
||||
|
||||
Ok(())
|
||||
}
|
||||
|
||||
pub async fn fetch_cache() -> AppState {
|
||||
let records = if let Ok(file) = tokio::fs::File::open(".cache/data").await.as_mut() {
|
||||
let mut buf: Vec<u8> = Vec::with_capacity(200);
|
||||
file.read_to_end(&mut buf).await.unwrap();
|
||||
|
||||
bincode::deserialize_from(&mut buf.as_slice()).unwrap()
|
||||
} else {
|
||||
HashMap::new()
|
||||
};
|
||||
|
||||
AppState::new(records)
|
||||
}
|
171
src/main.rs
171
src/main.rs
|
@ -1,45 +1,36 @@
|
|||
use async_zip::tokio::write::ZipFileWriter;
|
||||
use async_zip::{Compression, ZipEntryBuilder};
|
||||
use async_zip::{tokio::write::ZipFileWriter, Compression, ZipEntryBuilder};
|
||||
|
||||
use axum::body::StreamBody;
|
||||
use axum::extract::{ConnectInfo, State};
|
||||
use axum::http::{Request, StatusCode};
|
||||
use axum::middleware::{self, Next};
|
||||
use axum::response::{IntoResponse, Response};
|
||||
use axum::routing::{get, post};
|
||||
use axum::TypedHeader;
|
||||
use axum::{
|
||||
extract::{DefaultBodyLimit, Multipart},
|
||||
response::Redirect,
|
||||
Router,
|
||||
body::StreamBody,
|
||||
extract::{ConnectInfo, DefaultBodyLimit, Multipart, State},
|
||||
http::{Request, StatusCode},
|
||||
middleware::{self, Next},
|
||||
response::{IntoResponse, Redirect, Response},
|
||||
routing::{get, post},
|
||||
Router, TypedHeader,
|
||||
};
|
||||
|
||||
use futures::TryStreamExt;
|
||||
|
||||
use headers::{Header, HeaderName, HeaderValue};
|
||||
use rand::distributions::{Alphanumeric, DistString};
|
||||
use rand::rngs::SmallRng;
|
||||
use rand::SeedableRng;
|
||||
use nyazoom_headers::ForwardedFor;
|
||||
|
||||
use sanitize_filename_reader_friendly::sanitize;
|
||||
|
||||
use serde::Serialize;
|
||||
use std::{io, net::SocketAddr, path::Path};
|
||||
|
||||
use tokio::io::AsyncReadExt;
|
||||
use tokio_util::compat::FuturesAsyncWriteCompatExt;
|
||||
|
||||
use std::collections::HashMap;
|
||||
use std::io;
|
||||
use std::net::SocketAddr;
|
||||
use std::path::Path;
|
||||
|
||||
use tokio_util::io::{ReaderStream, StreamReader};
|
||||
use tokio_util::{
|
||||
compat::FuturesAsyncWriteCompatExt,
|
||||
io::{ReaderStream, StreamReader},
|
||||
};
|
||||
|
||||
use tower_http::{limit::RequestBodyLimitLayer, services::ServeDir, trace::TraceLayer};
|
||||
|
||||
use tracing_subscriber::{layer::SubscriberExt, util::SubscriberInitExt};
|
||||
|
||||
mod cache;
|
||||
mod nyazoom_headers;
|
||||
mod state;
|
||||
mod util;
|
||||
|
||||
use state::{AppState, UploadRecord};
|
||||
|
||||
|
@ -63,9 +54,9 @@ async fn main() -> io::Result<()> {
|
|||
.init();
|
||||
|
||||
// uses create_dir_all to create both .cache and serve inside it in one go
|
||||
make_dir(".cache/serve").await?;
|
||||
util::make_dir(".cache/serve").await?;
|
||||
|
||||
let state = fetch_cache().await;
|
||||
let state = cache::fetch_cache().await;
|
||||
|
||||
// Router Setup
|
||||
let app = Router::new()
|
||||
|
@ -91,23 +82,13 @@ async fn main() -> io::Result<()> {
|
|||
Ok(())
|
||||
}
|
||||
|
||||
// async fn log_source<B>(
|
||||
// ConnectInfo(addr): ConnectInfo<SocketAddr>,
|
||||
// req: Request<B>,
|
||||
// next: Next<B>,
|
||||
// ) -> Response {
|
||||
// tracing::info!("{}", addr);
|
||||
//
|
||||
// next.run(req).await
|
||||
// }
|
||||
|
||||
async fn log_source<B>(
|
||||
ConnectInfo(addr): ConnectInfo<SocketAddr>,
|
||||
TypedHeader(ForwardedFor(forwarded_for)): TypedHeader<ForwardedFor>,
|
||||
forwarded_for: Option<TypedHeader<ForwardedFor>>,
|
||||
req: Request<B>,
|
||||
next: Next<B>,
|
||||
) -> Response {
|
||||
tracing::info!("{} : {}", addr, forwarded_for);
|
||||
tracing::info!("{} : {:?}", addr, forwarded_for);
|
||||
|
||||
next.run(req).await
|
||||
}
|
||||
|
@ -118,7 +99,7 @@ async fn upload_to_zip(
|
|||
) -> Result<Redirect, (StatusCode, String)> {
|
||||
tracing::debug!("{:?}", *state.records.lock().await);
|
||||
|
||||
let cache_name = get_random_name(10);
|
||||
let cache_name = util::get_random_name(10);
|
||||
|
||||
let archive_path = Path::new(".cache/serve").join(&format!("{}.zip", &cache_name));
|
||||
|
||||
|
@ -162,7 +143,7 @@ async fn upload_to_zip(
|
|||
let mut records = state.records.lock().await;
|
||||
records.insert(cache_name.clone(), UploadRecord::new(archive_path));
|
||||
|
||||
write_to_cache(&records)
|
||||
cache::write_to_cache(&records)
|
||||
.await
|
||||
.map_err(|err| (StatusCode::INTERNAL_SERVER_ERROR, err.to_string()))?;
|
||||
|
||||
|
@ -189,113 +170,11 @@ async fn download(
|
|||
.unwrap()
|
||||
.into_response());
|
||||
} else {
|
||||
let _ = tokio::fs::remove_file(&record.file);
|
||||
let _ = tokio::fs::remove_file(&record.file).await;
|
||||
records.remove(&id);
|
||||
write_to_cache(&records).await.unwrap();
|
||||
cache::write_to_cache(&records).await.unwrap();
|
||||
}
|
||||
}
|
||||
|
||||
Ok(Redirect::to("/404.html").into_response())
|
||||
}
|
||||
|
||||
#[inline]
|
||||
async fn make_dir<T>(name: T) -> io::Result<()>
|
||||
where
|
||||
T: AsRef<Path>,
|
||||
{
|
||||
tokio::fs::create_dir_all(name)
|
||||
.await
|
||||
.or_else(|err| match err.kind() {
|
||||
io::ErrorKind::AlreadyExists => Ok(()),
|
||||
_ => Err(err),
|
||||
})
|
||||
}
|
||||
|
||||
#[inline]
|
||||
fn get_random_name(len: usize) -> String {
|
||||
let mut rng = SmallRng::from_entropy();
|
||||
|
||||
Alphanumeric.sample_string(&mut rng, len)
|
||||
}
|
||||
|
||||
async fn write_to_cache<T, Y>(records: &HashMap<T, Y>) -> io::Result<()>
|
||||
where
|
||||
T: Serialize,
|
||||
Y: Serialize,
|
||||
{
|
||||
let mut records_cache = tokio::fs::File::create(".cache/data").await.unwrap();
|
||||
|
||||
let mut buf: Vec<u8> = Vec::with_capacity(200);
|
||||
bincode::serialize_into(&mut buf, &*records)
|
||||
.map_err(|err| error::io_other(&err.to_string()))?;
|
||||
|
||||
let bytes_written = tokio::io::copy(&mut buf.as_slice(), &mut records_cache).await?;
|
||||
|
||||
tracing::debug!("state cache size: {}", bytes_written);
|
||||
|
||||
Ok(())
|
||||
}
|
||||
|
||||
async fn fetch_cache() -> AppState {
|
||||
let records = if let Ok(file) = tokio::fs::File::open(".cache/data").await.as_mut() {
|
||||
let mut buf: Vec<u8> = Vec::with_capacity(200);
|
||||
file.read_to_end(&mut buf).await.unwrap();
|
||||
|
||||
bincode::deserialize_from(&mut buf.as_slice()).unwrap()
|
||||
} else {
|
||||
HashMap::new()
|
||||
};
|
||||
|
||||
AppState::new(records)
|
||||
}
|
||||
|
||||
#[allow(dead_code)]
|
||||
static UNITS: [&str; 6] = ["KiB", "MiB", "GiB", "TiB", "PiB", "EiB"];
|
||||
// This function is actually rather interesting to me, I understand that rust is
|
||||
// very powerful, and its very safe, but i find it rather amusing that the [] operator
|
||||
// doesn't check bounds, meaning it can panic at runtime. Usually rust is very
|
||||
// very careful about possible panics
|
||||
//
|
||||
// although this function shouldn't be able to panic at runtime due to known bounds
|
||||
// being listened to
|
||||
#[inline]
|
||||
fn _bytes_to_human_readable(bytes: u64) -> String {
|
||||
let mut running = bytes as f64;
|
||||
let mut count = 0;
|
||||
while running > 1024.0 && count <= 6 {
|
||||
running /= 1024.0;
|
||||
count += 1;
|
||||
}
|
||||
|
||||
format!("{:.2} {}", running, UNITS[count - 1])
|
||||
}
|
||||
|
||||
struct ForwardedFor(String);
|
||||
|
||||
static FF_TEXT: &str = "x-forwarded-for";
|
||||
static FF_NAME: HeaderName = HeaderName::from_static(FF_TEXT);
|
||||
|
||||
impl Header for ForwardedFor {
|
||||
fn name() -> &'static HeaderName {
|
||||
&FF_NAME
|
||||
}
|
||||
|
||||
fn decode<'i, I>(values: &mut I) -> Result<Self, headers::Error>
|
||||
where
|
||||
Self: Sized,
|
||||
I: Iterator<Item = &'i headers::HeaderValue>,
|
||||
{
|
||||
let value = values
|
||||
.next()
|
||||
.ok_or_else(headers::Error::invalid)?
|
||||
.to_str()
|
||||
.map_err(|_| headers::Error::invalid())?
|
||||
.to_owned();
|
||||
|
||||
Ok(ForwardedFor(value))
|
||||
}
|
||||
|
||||
fn encode<E: Extend<headers::HeaderValue>>(&self, values: &mut E) {
|
||||
values.extend(std::iter::once(HeaderValue::from_str(&self.0).unwrap()));
|
||||
}
|
||||
}
|
||||
|
|
33
src/nyazoom_headers.rs
Normal file
33
src/nyazoom_headers.rs
Normal file
|
@ -0,0 +1,33 @@
|
|||
use headers::{self, Header, HeaderName, HeaderValue};
|
||||
|
||||
#[derive(Debug)]
|
||||
pub struct ForwardedFor(String);
|
||||
|
||||
pub static FF_TEXT: &str = "x-forwarded-for";
|
||||
|
||||
pub static FF_NAME: HeaderName = HeaderName::from_static(FF_TEXT);
|
||||
|
||||
impl Header for ForwardedFor {
|
||||
fn name() -> &'static HeaderName {
|
||||
&FF_NAME
|
||||
}
|
||||
|
||||
fn decode<'i, I>(values: &mut I) -> Result<Self, headers::Error>
|
||||
where
|
||||
Self: Sized,
|
||||
I: Iterator<Item = &'i headers::HeaderValue>,
|
||||
{
|
||||
let value = values
|
||||
.next()
|
||||
.ok_or_else(headers::Error::invalid)?
|
||||
.to_str()
|
||||
.map_err(|_| headers::Error::invalid())?
|
||||
.to_owned();
|
||||
|
||||
Ok(ForwardedFor(value))
|
||||
}
|
||||
|
||||
fn encode<E: Extend<headers::HeaderValue>>(&self, values: &mut E) {
|
||||
values.extend(std::iter::once(HeaderValue::from_str(&self.0).unwrap()));
|
||||
}
|
||||
}
|
49
src/util.rs
Normal file
49
src/util.rs
Normal file
|
@ -0,0 +1,49 @@
|
|||
use rand::{
|
||||
distributions::{Alphanumeric, DistString},
|
||||
rngs::SmallRng,
|
||||
SeedableRng,
|
||||
};
|
||||
|
||||
use std::{io, path::Path};
|
||||
|
||||
#[inline]
|
||||
pub async fn make_dir<T>(name: T) -> io::Result<()>
|
||||
where
|
||||
T: AsRef<Path>,
|
||||
{
|
||||
tokio::fs::create_dir_all(name)
|
||||
.await
|
||||
.or_else(|err| match err.kind() {
|
||||
io::ErrorKind::AlreadyExists => Ok(()),
|
||||
_ => Err(err),
|
||||
})
|
||||
}
|
||||
|
||||
#[inline]
|
||||
pub fn get_random_name(len: usize) -> String {
|
||||
let mut rng = SmallRng::from_entropy();
|
||||
|
||||
Alphanumeric.sample_string(&mut rng, len)
|
||||
}
|
||||
|
||||
#[allow(dead_code)]
|
||||
pub static UNITS: [&str; 6] = ["KiB", "MiB", "GiB", "TiB", "PiB", "EiB"];
|
||||
|
||||
// This function is actually rather interesting to me, I understand that rust is
|
||||
// very powerful, and its very safe, but i find it rather amusing that the [] operator
|
||||
// doesn't check bounds, meaning it can panic at runtime. Usually rust is very
|
||||
// very careful about possible panics
|
||||
//
|
||||
// although this function shouldn't be able to panic at runtime due to known bounds
|
||||
// being listened to
|
||||
#[inline]
|
||||
pub fn _bytes_to_human_readable(bytes: u64) -> String {
|
||||
let mut running = bytes as f64;
|
||||
let mut count = 0;
|
||||
while running > 1024.0 && count <= 6 {
|
||||
running /= 1024.0;
|
||||
count += 1;
|
||||
}
|
||||
|
||||
format!("{:.2} {}", running, UNITS[count - 1])
|
||||
}
|
Loading…
Reference in a new issue