Added MultipartFile extractors.
Moved cfg macro to lib where possible. Changed some features, and made some deps optional
This commit is contained in:
parent
e0baff8625
commit
0898a50166
62
Cargo.lock
generated
62
Cargo.lock
generated
@ -47,6 +47,7 @@ dependencies = [
|
|||||||
"matchit",
|
"matchit",
|
||||||
"memchr",
|
"memchr",
|
||||||
"mime",
|
"mime",
|
||||||
|
"multer",
|
||||||
"percent-encoding",
|
"percent-encoding",
|
||||||
"pin-project-lite",
|
"pin-project-lite",
|
||||||
"rustversion",
|
"rustversion",
|
||||||
@ -130,6 +131,15 @@ dependencies = [
|
|||||||
"syn",
|
"syn",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "encoding_rs"
|
||||||
|
version = "0.8.34"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "b45de904aa0b010bce2ab45264d0631681847fa7b6f2eaa7dab7619943bc4f59"
|
||||||
|
dependencies = [
|
||||||
|
"cfg-if",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "fnv"
|
name = "fnv"
|
||||||
version = "1.0.7"
|
version = "1.0.7"
|
||||||
@ -284,12 +294,13 @@ checksum = "bbd2bcb4c963f2ddae06a2efc7e9f3591312473c50c6685e1f298068316e66fe"
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "lib"
|
name = "lib"
|
||||||
version = "1.1.1"
|
version = "1.3.0"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"axum",
|
"axum",
|
||||||
"derive",
|
"derive",
|
||||||
"nom",
|
"nom",
|
||||||
"serde",
|
"serde",
|
||||||
|
"thiserror",
|
||||||
"tokio",
|
"tokio",
|
||||||
"tokio-util",
|
"tokio-util",
|
||||||
"tower",
|
"tower",
|
||||||
@ -354,6 +365,23 @@ dependencies = [
|
|||||||
"windows-sys 0.48.0",
|
"windows-sys 0.48.0",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "multer"
|
||||||
|
version = "3.1.0"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "83e87776546dc87511aa5ee218730c92b666d7264ab6ed41f9d215af9cd5224b"
|
||||||
|
dependencies = [
|
||||||
|
"bytes",
|
||||||
|
"encoding_rs",
|
||||||
|
"futures-util",
|
||||||
|
"http",
|
||||||
|
"httparse",
|
||||||
|
"memchr",
|
||||||
|
"mime",
|
||||||
|
"spin",
|
||||||
|
"version_check",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "nom"
|
name = "nom"
|
||||||
version = "7.1.3"
|
version = "7.1.3"
|
||||||
@ -547,6 +575,12 @@ dependencies = [
|
|||||||
"windows-sys 0.52.0",
|
"windows-sys 0.52.0",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "spin"
|
||||||
|
version = "0.9.8"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "6980e8d7511241f8acf4aebddbb1ff938df5eebe98691418c4468d0b72a96a67"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "syn"
|
name = "syn"
|
||||||
version = "2.0.67"
|
version = "2.0.67"
|
||||||
@ -570,6 +604,26 @@ version = "1.0.1"
|
|||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "a7065abeca94b6a8a577f9bd45aa0867a2238b74e8eb67cf10d492bc39351394"
|
checksum = "a7065abeca94b6a8a577f9bd45aa0867a2238b74e8eb67cf10d492bc39351394"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "thiserror"
|
||||||
|
version = "1.0.61"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "c546c80d6be4bc6a00c0f01730c08df82eaa7a7a61f11d656526506112cc1709"
|
||||||
|
dependencies = [
|
||||||
|
"thiserror-impl",
|
||||||
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "thiserror-impl"
|
||||||
|
version = "1.0.61"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "46c3384250002a6d5af4d114f2845d37b57521033f30d5c3f46c4d70e1197533"
|
||||||
|
dependencies = [
|
||||||
|
"proc-macro2",
|
||||||
|
"quote",
|
||||||
|
"syn",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "thread_local"
|
name = "thread_local"
|
||||||
version = "1.1.8"
|
version = "1.1.8"
|
||||||
@ -734,6 +788,12 @@ version = "0.1.0"
|
|||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "830b7e5d4d90034032940e4ace0d9a9a057e7a45cd94e6c007832e39edb82f6d"
|
checksum = "830b7e5d4d90034032940e4ace0d9a9a057e7a45cd94e6c007832e39edb82f6d"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "version_check"
|
||||||
|
version = "0.9.4"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "49874b5167b65d7193b8aba1567f5c7d93d001cafc34600cee003eda787e483f"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "wasi"
|
name = "wasi"
|
||||||
version = "0.11.0+wasi-snapshot-preview1"
|
version = "0.11.0+wasi-snapshot-preview1"
|
||||||
|
17
Cargo.toml
17
Cargo.toml
@ -1,6 +1,6 @@
|
|||||||
[package]
|
[package]
|
||||||
name = "lib"
|
name = "lib"
|
||||||
version = "1.1.1"
|
version = "1.3.0"
|
||||||
edition = "2021"
|
edition = "2021"
|
||||||
authors = ["Martin Berg Alstad"]
|
authors = ["Martin Berg Alstad"]
|
||||||
|
|
||||||
@ -8,15 +8,17 @@ authors = ["Martin Berg Alstad"]
|
|||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
# Api
|
# Api
|
||||||
axum = { version = "0.7.5", optional = true }
|
axum = { version = "0.7.5", optional = true, features = ["multipart"] }
|
||||||
tower = { version = "0.4.13", optional = true }
|
tower = { version = "0.4.13", optional = true }
|
||||||
tower-http = { version = "0.5.2", optional = true, features = ["trace", "cors", "normalize-path"] }
|
tower-http = { version = "0.5.2", optional = true, features = ["trace", "cors", "normalize-path"] }
|
||||||
# Async
|
# Async
|
||||||
tokio = { version = "1.38.0", optional = true, features = ["fs"] }
|
tokio = { version = "1.38.0", optional = true, features = ["fs"] }
|
||||||
tokio-util = { version = "0.7.11", optional = true, features = ["io"] }
|
tokio-util = { version = "0.7.11", optional = true, features = ["io"] }
|
||||||
|
# Error handling
|
||||||
|
thiserror = { version = "1.0.61", optional = true }
|
||||||
# Logging
|
# Logging
|
||||||
tracing = "0.1.40"
|
tracing = { version = "0.1.40", optional = true }
|
||||||
tracing-subscriber = "0.3.18"
|
tracing-subscriber = { version = "0.3.18", optional = true }
|
||||||
# Parsing
|
# Parsing
|
||||||
nom = { version = "7.1.3", optional = true }
|
nom = { version = "7.1.3", optional = true }
|
||||||
# Serialization / Deserialization
|
# Serialization / Deserialization
|
||||||
@ -25,9 +27,10 @@ serde = { version = "1.0.203", optional = true, features = ["derive"] }
|
|||||||
derive = { path = "derive", optional = true }
|
derive = { path = "derive", optional = true }
|
||||||
|
|
||||||
[features]
|
[features]
|
||||||
axum = ["dep:axum", "dep:tower", "dep:tower-http"]
|
axum = ["dep:axum", "dep:tower", "dep:tower-http", "dep:thiserror", "dep:tracing", "dep:tracing-subscriber"]
|
||||||
tokio = ["dep:tokio", "dep:tokio-util"]
|
tokio = ["dep:tokio"]
|
||||||
vec = []
|
io = ["dep:tokio", "dep:tokio-util"]
|
||||||
|
iter = []
|
||||||
nom = ["dep:nom"]
|
nom = ["dep:nom"]
|
||||||
serde = ["dep:serde"]
|
serde = ["dep:serde"]
|
||||||
derive = ["dep:derive", "axum", "serde"]
|
derive = ["dep:derive", "axum", "serde"]
|
||||||
|
@ -1,4 +1,3 @@
|
|||||||
#[cfg(feature = "axum")]
|
|
||||||
use {
|
use {
|
||||||
axum::{extract::Request, handler::Handler, Router, ServiceExt},
|
axum::{extract::Request, handler::Handler, Router, ServiceExt},
|
||||||
std::net::Ipv4Addr,
|
std::net::Ipv4Addr,
|
||||||
@ -11,12 +10,11 @@ use {
|
|||||||
},
|
},
|
||||||
tracing::{info, Level},
|
tracing::{info, Level},
|
||||||
};
|
};
|
||||||
#[cfg(all(feature = "axum", feature = "tokio"))]
|
#[cfg(feature = "tokio")]
|
||||||
use {std::io, std::net::SocketAddr, tokio::net::TcpListener};
|
use {std::io, std::net::SocketAddr, tokio::net::TcpListener};
|
||||||
|
|
||||||
// TODO trim trailing slash into macro > let _app = NormalizePathLayer::trim_trailing_slash().layer(create_app!(routes));
|
// TODO trim trailing slash into macro > let _app = NormalizePathLayer::trim_trailing_slash().layer(create_app!(routes));
|
||||||
#[macro_export]
|
#[macro_export]
|
||||||
#[cfg(feature = "axum")]
|
|
||||||
macro_rules! create_app {
|
macro_rules! create_app {
|
||||||
($router:expr) => {
|
($router:expr) => {
|
||||||
$router
|
$router
|
||||||
@ -27,7 +25,6 @@ macro_rules! create_app {
|
|||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Default)]
|
#[derive(Default)]
|
||||||
#[cfg(feature = "axum")]
|
|
||||||
pub struct AppBuilder {
|
pub struct AppBuilder {
|
||||||
router: Router,
|
router: Router,
|
||||||
socket: Option<(Ipv4Addr, u16)>,
|
socket: Option<(Ipv4Addr, u16)>,
|
||||||
@ -36,7 +33,6 @@ pub struct AppBuilder {
|
|||||||
tracing: Option<TraceLayer<HttpMakeClassifier>>,
|
tracing: Option<TraceLayer<HttpMakeClassifier>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(all(feature = "axum", feature = "tokio"))]
|
|
||||||
impl AppBuilder {
|
impl AppBuilder {
|
||||||
pub fn new() -> Self {
|
pub fn new() -> Self {
|
||||||
Self::default()
|
Self::default()
|
||||||
@ -81,8 +77,9 @@ impl AppBuilder {
|
|||||||
self
|
self
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[cfg(feature = "tokio")]
|
||||||
pub async fn serve(self) -> io::Result<()> {
|
pub async fn serve(self) -> io::Result<()> {
|
||||||
let _ = fmt_trace();
|
let _ = fmt_trace(); // Allowed to fail
|
||||||
let listener = self.listener().await?;
|
let listener = self.listener().await?;
|
||||||
|
|
||||||
if self.normalize_path.unwrap_or(true) {
|
if self.normalize_path.unwrap_or(true) {
|
||||||
@ -95,6 +92,7 @@ impl AppBuilder {
|
|||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[cfg(feature = "tokio")]
|
||||||
async fn listener(&self) -> io::Result<TcpListener> {
|
async fn listener(&self) -> io::Result<TcpListener> {
|
||||||
let addr = SocketAddr::from(self.socket.unwrap_or((Ipv4Addr::UNSPECIFIED, 8000)));
|
let addr = SocketAddr::from(self.socket.unwrap_or((Ipv4Addr::UNSPECIFIED, 8000)));
|
||||||
info!("Initializing server on: {addr}");
|
info!("Initializing server on: {addr}");
|
||||||
@ -124,7 +122,7 @@ fn fmt_trace() -> Result<(), String> {
|
|||||||
.map_err(|error| error.to_string())
|
.map_err(|error| error.to_string())
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(all(test, feature = "axum"))]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use axum::Router;
|
use axum::Router;
|
||||||
|
|
||||||
|
187
src/axum/extractor.rs
Normal file
187
src/axum/extractor.rs
Normal file
@ -0,0 +1,187 @@
|
|||||||
|
use axum::{
|
||||||
|
async_trait,
|
||||||
|
extract::{
|
||||||
|
multipart::{Field, MultipartError, MultipartRejection},
|
||||||
|
FromRequest, Multipart, Request,
|
||||||
|
},
|
||||||
|
response::IntoResponse,
|
||||||
|
};
|
||||||
|
use thiserror::Error;
|
||||||
|
|
||||||
|
#[derive(PartialEq, Eq, Ord, PartialOrd, Hash, Debug, Clone, Copy)]
|
||||||
|
pub enum ContentType {
|
||||||
|
Json,
|
||||||
|
Form,
|
||||||
|
Multipart,
|
||||||
|
Pdf,
|
||||||
|
Html,
|
||||||
|
Unknown,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl From<&str> for ContentType {
|
||||||
|
fn from(content_type: &str) -> Self {
|
||||||
|
match content_type {
|
||||||
|
"application/json" => ContentType::Json,
|
||||||
|
"application/x-www-form-urlencoded" => ContentType::Form,
|
||||||
|
"multipart/form-data" => ContentType::Multipart,
|
||||||
|
"application/pdf" => ContentType::Pdf,
|
||||||
|
"text/html" => ContentType::Html,
|
||||||
|
_ => ContentType::Unknown,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl From<String> for ContentType {
|
||||||
|
fn from(content_type: String) -> Self {
|
||||||
|
ContentType::from(content_type.as_str())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl From<Option<&str>> for ContentType {
|
||||||
|
fn from(content_type: Option<&str>) -> Self {
|
||||||
|
content_type
|
||||||
|
.map(ContentType::from)
|
||||||
|
.unwrap_or(ContentType::Unknown)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub struct File {
|
||||||
|
pub filename: String,
|
||||||
|
pub bytes: Vec<u8>,
|
||||||
|
pub content_type: ContentType,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl File {
|
||||||
|
pub fn new(
|
||||||
|
filename: impl Into<String>,
|
||||||
|
bytes: impl Into<Vec<u8>>,
|
||||||
|
content_type: impl Into<ContentType>,
|
||||||
|
) -> Self {
|
||||||
|
Self {
|
||||||
|
filename: filename.into(),
|
||||||
|
bytes: bytes.into(),
|
||||||
|
content_type: content_type.into(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn from_field(field: Field<'_>) -> Result<Self, MultipartFileRejection> {
|
||||||
|
let filename = field
|
||||||
|
.file_name()
|
||||||
|
.ok_or(MultipartFileRejection::MissingFilename)?
|
||||||
|
.to_string();
|
||||||
|
let content_type: ContentType = field.content_type().into();
|
||||||
|
let bytes = field.bytes().await?;
|
||||||
|
Ok(File::new(filename, bytes, content_type))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Extractor for a single file from a multipart request.
|
||||||
|
/// Expects exactly one file. A file must have a name, bytes and optionally a content type.
|
||||||
|
/// This extractor consumes the request and must ble placed last in the handler.
|
||||||
|
pub struct MultipartFile(pub File);
|
||||||
|
/// Extractor for multiple files from a multipart request.
|
||||||
|
/// Expects at least one file. A file must have a name, bytes and optionally a content type.
|
||||||
|
/// This extractor consumes the request and must ble placed last in the handler.
|
||||||
|
pub struct MultipartFiles(pub Vec<File>);
|
||||||
|
|
||||||
|
#[derive(Debug, Error)]
|
||||||
|
pub enum MultipartFileRejection {
|
||||||
|
#[error(transparent)]
|
||||||
|
MultipartRejection(#[from] MultipartRejection),
|
||||||
|
#[error("Field error: {0}")]
|
||||||
|
FieldError(String),
|
||||||
|
#[error("No files found")]
|
||||||
|
NoFiles,
|
||||||
|
#[error("Expected one file, got several")]
|
||||||
|
SeveralFiles,
|
||||||
|
#[error("Missing filename")]
|
||||||
|
MissingFilename,
|
||||||
|
#[error("Error in body of multipart: {0}")]
|
||||||
|
BodyError(String),
|
||||||
|
}
|
||||||
|
|
||||||
|
impl From<MultipartError> for MultipartFileRejection {
|
||||||
|
fn from(error: MultipartError) -> Self {
|
||||||
|
MultipartFileRejection::BodyError(error.body_text())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl IntoResponse for MultipartFileRejection {
|
||||||
|
fn into_response(self) -> axum::response::Response {
|
||||||
|
match self {
|
||||||
|
MultipartFileRejection::MultipartRejection(rejection) => rejection.into_response(),
|
||||||
|
MultipartFileRejection::FieldError(error) => {
|
||||||
|
(axum::http::StatusCode::BAD_REQUEST, error).into_response()
|
||||||
|
}
|
||||||
|
MultipartFileRejection::NoFiles => {
|
||||||
|
(axum::http::StatusCode::BAD_REQUEST, "No files found").into_response()
|
||||||
|
}
|
||||||
|
MultipartFileRejection::SeveralFiles => (
|
||||||
|
axum::http::StatusCode::BAD_REQUEST,
|
||||||
|
"Expected one file, got several",
|
||||||
|
)
|
||||||
|
.into_response(),
|
||||||
|
MultipartFileRejection::MissingFilename => {
|
||||||
|
(axum::http::StatusCode::BAD_REQUEST, "Missing filename").into_response()
|
||||||
|
}
|
||||||
|
MultipartFileRejection::BodyError(error) => {
|
||||||
|
(axum::http::StatusCode::BAD_REQUEST, error).into_response()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[async_trait]
|
||||||
|
impl<S> FromRequest<S> for MultipartFile
|
||||||
|
where
|
||||||
|
S: Send + Sync,
|
||||||
|
{
|
||||||
|
type Rejection = MultipartFileRejection;
|
||||||
|
|
||||||
|
async fn from_request(req: Request, state: &S) -> Result<Self, Self::Rejection> {
|
||||||
|
let mut multipart = Multipart::from_request(req, state).await?;
|
||||||
|
let fields = get_fields(&mut multipart).await?;
|
||||||
|
if fields.len() > 1 {
|
||||||
|
Err(MultipartFileRejection::SeveralFiles)
|
||||||
|
} else {
|
||||||
|
let field = fields
|
||||||
|
.into_iter()
|
||||||
|
.next()
|
||||||
|
.ok_or(MultipartFileRejection::NoFiles)?;
|
||||||
|
Ok(MultipartFile(File::from_field(field).await?))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[async_trait]
|
||||||
|
impl<S> FromRequest<S> for MultipartFiles
|
||||||
|
where
|
||||||
|
S: Send + Sync,
|
||||||
|
{
|
||||||
|
type Rejection = MultipartFileRejection;
|
||||||
|
|
||||||
|
async fn from_request(req: Request, state: &S) -> Result<Self, Self::Rejection> {
|
||||||
|
let mut multipart = Multipart::from_request(req, state).await?;
|
||||||
|
let fields = get_fields(&mut multipart).await?;
|
||||||
|
if fields.is_empty() {
|
||||||
|
Err(MultipartFileRejection::NoFiles)
|
||||||
|
} else {
|
||||||
|
let mut files = vec![];
|
||||||
|
for field in fields.into_iter() {
|
||||||
|
files.push(File::from_field(field).await?);
|
||||||
|
}
|
||||||
|
Ok(MultipartFiles(files))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn get_fields<'a>(
|
||||||
|
multipart: &'a mut Multipart,
|
||||||
|
) -> Result<Vec<Field<'a>>, MultipartFileRejection> {
|
||||||
|
let fields: Vec<Field> = multipart.next_field().await?.into_iter().collect();
|
||||||
|
if fields.is_empty() {
|
||||||
|
Err(MultipartFileRejection::NoFiles)
|
||||||
|
} else {
|
||||||
|
Ok(fields)
|
||||||
|
}
|
||||||
|
}
|
@ -1,4 +1,4 @@
|
|||||||
#[cfg(all(feature = "tokio", feature = "axum"))]
|
#[cfg(feature = "tokio")]
|
||||||
use {crate::io::file, axum::body::Body, axum::response::Html, std::io};
|
use {crate::io::file, axum::body::Body, axum::response::Html, std::io};
|
||||||
|
|
||||||
/// Load an HTML file from the given file path, relative to the current directory.
|
/// Load an HTML file from the given file path, relative to the current directory.
|
||||||
@ -10,7 +10,7 @@ use {crate::io::file, axum::body::Body, axum::response::Html, std::io};
|
|||||||
/// ```
|
/// ```
|
||||||
/// let html = async { lib::axum::load::load_html("openapi.html").await.unwrap() };
|
/// let html = async { lib::axum::load::load_html("openapi.html").await.unwrap() };
|
||||||
/// ```
|
/// ```
|
||||||
#[cfg(all(feature = "tokio", feature = "axum"))]
|
#[cfg(feature = "tokio")]
|
||||||
pub async fn load_html<Path>(file_path: Path) -> Result<Html<Body>, io::Error>
|
pub async fn load_html<Path>(file_path: Path) -> Result<Html<Body>, io::Error>
|
||||||
where
|
where
|
||||||
Path: AsRef<std::path::Path>,
|
Path: AsRef<std::path::Path>,
|
||||||
@ -18,7 +18,7 @@ where
|
|||||||
load_file(file_path).await.map(Html)
|
load_file(file_path).await.map(Html)
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(all(feature = "tokio", feature = "axum"))]
|
#[cfg(feature = "tokio")]
|
||||||
pub async fn load_file<Path>(file_path: Path) -> Result<Body, io::Error>
|
pub async fn load_file<Path>(file_path: Path) -> Result<Body, io::Error>
|
||||||
where
|
where
|
||||||
Path: AsRef<std::path::Path>,
|
Path: AsRef<std::path::Path>,
|
||||||
@ -38,7 +38,6 @@ where
|
|||||||
/// ```
|
/// ```
|
||||||
// TODO check platform and use correct path separator
|
// TODO check platform and use correct path separator
|
||||||
#[macro_export]
|
#[macro_export]
|
||||||
#[cfg(feature = "axum")]
|
|
||||||
macro_rules! load_html {
|
macro_rules! load_html {
|
||||||
($filepath:expr) => {
|
($filepath:expr) => {
|
||||||
axum::response::Html(
|
axum::response::Html(
|
||||||
@ -58,7 +57,7 @@ macro_rules! load_html {
|
|||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(all(test, feature = "axum"))]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
#[test]
|
#[test]
|
||||||
fn test_load_html() {
|
fn test_load_html() {
|
||||||
|
@ -1,4 +1,6 @@
|
|||||||
pub mod app;
|
pub mod app;
|
||||||
|
pub mod extractor;
|
||||||
pub mod load;
|
pub mod load;
|
||||||
|
#[cfg(feature = "serde")]
|
||||||
pub mod response;
|
pub mod response;
|
||||||
pub mod router;
|
pub mod router;
|
||||||
|
@ -1,4 +1,3 @@
|
|||||||
#[cfg(all(feature = "axum", feature = "serde"))]
|
|
||||||
use {
|
use {
|
||||||
crate::serde::response::BaseResponse,
|
crate::serde::response::BaseResponse,
|
||||||
axum::{
|
axum::{
|
||||||
@ -8,14 +7,13 @@ use {
|
|||||||
serde::Serialize,
|
serde::Serialize,
|
||||||
};
|
};
|
||||||
|
|
||||||
#[cfg(all(feature = "axum", feature = "serde"))]
|
|
||||||
impl<T: Serialize> IntoResponse for BaseResponse<T> {
|
impl<T: Serialize> IntoResponse for BaseResponse<T> {
|
||||||
fn into_response(self) -> Response {
|
fn into_response(self) -> Response {
|
||||||
Json(self).into_response()
|
Json(self).into_response()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(all(test, feature = "axum", feature = "serde"))]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use axum::http::header::CONTENT_TYPE;
|
use axum::http::header::CONTENT_TYPE;
|
||||||
use axum::http::{HeaderValue, StatusCode};
|
use axum::http::{HeaderValue, StatusCode};
|
||||||
|
@ -18,7 +18,6 @@
|
|||||||
/// ));
|
/// ));
|
||||||
/// ```
|
/// ```
|
||||||
#[macro_export]
|
#[macro_export]
|
||||||
#[cfg(feature = "axum")]
|
|
||||||
macro_rules! router {
|
macro_rules! router {
|
||||||
($body:expr) => {
|
($body:expr) => {
|
||||||
pub(crate) fn router() -> axum::Router {
|
pub(crate) fn router() -> axum::Router {
|
||||||
@ -52,7 +51,6 @@ macro_rules! router {
|
|||||||
/// );
|
/// );
|
||||||
/// ```
|
/// ```
|
||||||
#[macro_export]
|
#[macro_export]
|
||||||
#[cfg(feature = "axum")]
|
|
||||||
macro_rules! routes {
|
macro_rules! routes {
|
||||||
($($method:ident $route:expr => $func:expr),* $(,)?) => {
|
($($method:ident $route:expr => $func:expr),* $(,)?) => {
|
||||||
axum::Router::new()
|
axum::Router::new()
|
||||||
@ -63,7 +61,6 @@ macro_rules! routes {
|
|||||||
}
|
}
|
||||||
|
|
||||||
#[macro_export]
|
#[macro_export]
|
||||||
#[cfg(feature = "axum")]
|
|
||||||
macro_rules! join_routes {
|
macro_rules! join_routes {
|
||||||
($($route:expr),* $(,)?) => {
|
($($route:expr),* $(,)?) => {
|
||||||
axum::Router::new()$(
|
axum::Router::new()$(
|
||||||
@ -72,7 +69,7 @@ macro_rules! join_routes {
|
|||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(all(test, feature = "axum"))]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use axum::extract::State;
|
use axum::extract::State;
|
||||||
use axum::Router;
|
use axum::Router;
|
||||||
@ -117,7 +114,7 @@ mod tests {
|
|||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_routes() {
|
fn test_routes() {
|
||||||
let _router: Router<()> = routes!(
|
let _router: Router = routes!(
|
||||||
get "/" => index,
|
get "/" => index,
|
||||||
post "/" => || async {}
|
post "/" => || async {}
|
||||||
);
|
);
|
||||||
|
@ -1,7 +1,5 @@
|
|||||||
#[cfg(feature = "tokio")]
|
|
||||||
use {std::io::Error, tokio::fs::File, tokio_util::io::ReaderStream};
|
use {std::io::Error, tokio::fs::File, tokio_util::io::ReaderStream};
|
||||||
|
|
||||||
#[cfg(feature = "tokio")]
|
|
||||||
pub async fn load_file<Path>(file_path: Path) -> Result<ReaderStream<File>, Error>
|
pub async fn load_file<Path>(file_path: Path) -> Result<ReaderStream<File>, Error>
|
||||||
where
|
where
|
||||||
Path: AsRef<std::path::Path>,
|
Path: AsRef<std::path::Path>,
|
||||||
@ -9,7 +7,7 @@ where
|
|||||||
File::open(file_path).await.map(ReaderStream::new)
|
File::open(file_path).await.map(ReaderStream::new)
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(all(test, feature = "tokio"))]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use super::*;
|
use super::*;
|
||||||
|
|
||||||
|
@ -1,11 +1,16 @@
|
|||||||
#![allow(dead_code)]
|
#![allow(dead_code)]
|
||||||
|
|
||||||
|
#[cfg(feature = "axum")]
|
||||||
pub mod axum;
|
pub mod axum;
|
||||||
|
#[cfg(feature = "io")]
|
||||||
pub mod io;
|
pub mod io;
|
||||||
|
#[cfg(feature = "nom")]
|
||||||
pub mod nom;
|
pub mod nom;
|
||||||
|
#[cfg(feature = "serde")]
|
||||||
pub mod serde;
|
pub mod serde;
|
||||||
mod traits;
|
pub mod traits;
|
||||||
|
#[cfg(feature = "iter")]
|
||||||
pub mod vector;
|
pub mod vector;
|
||||||
|
|
||||||
#[cfg(feature = "derive")]
|
#[cfg(all(feature = "derive", feature = "serde"))]
|
||||||
pub extern crate derive;
|
pub extern crate derive;
|
||||||
|
@ -1,4 +1,3 @@
|
|||||||
#[cfg(feature = "nom")]
|
|
||||||
use {
|
use {
|
||||||
nom::{
|
nom::{
|
||||||
bytes::complete::take_while_m_n,
|
bytes::complete::take_while_m_n,
|
||||||
@ -16,7 +15,6 @@ use {
|
|||||||
/// - Parameters
|
/// - Parameters
|
||||||
/// - `inner`: The parser to trim
|
/// - `inner`: The parser to trim
|
||||||
/// - Returns: A parser that trims leading and trailing whitespace from the input and then runs the value from the inner parser
|
/// - Returns: A parser that trims leading and trailing whitespace from the input and then runs the value from the inner parser
|
||||||
#[cfg(feature = "nom")]
|
|
||||||
pub fn trim<'a, Parser, R>(inner: Parser) -> impl FnMut(&'a str) -> IResult<&'a str, R>
|
pub fn trim<'a, Parser, R>(inner: Parser) -> impl FnMut(&'a str) -> IResult<&'a str, R>
|
||||||
where
|
where
|
||||||
Parser: Fn(&'a str) -> IResult<&'a str, R>,
|
Parser: Fn(&'a str) -> IResult<&'a str, R>,
|
||||||
@ -29,7 +27,6 @@ where
|
|||||||
/// - Parameters
|
/// - Parameters
|
||||||
/// - `inner`: The parser to run inside the parentheses
|
/// - `inner`: The parser to run inside the parentheses
|
||||||
/// - Returns: A parser that parses a parenthesized expression
|
/// - Returns: A parser that parses a parenthesized expression
|
||||||
#[cfg(feature = "nom")]
|
|
||||||
pub fn parenthesized<'a, Parser, R>(inner: Parser) -> impl FnMut(&'a str) -> IResult<&'a str, R>
|
pub fn parenthesized<'a, Parser, R>(inner: Parser) -> impl FnMut(&'a str) -> IResult<&'a str, R>
|
||||||
where
|
where
|
||||||
Parser: Fn(&'a str) -> IResult<&'a str, R>,
|
Parser: Fn(&'a str) -> IResult<&'a str, R>,
|
||||||
@ -42,7 +39,6 @@ where
|
|||||||
/// - `n`: The length of the string to take
|
/// - `n`: The length of the string to take
|
||||||
/// - `predicate`: The predicate to call to validate the input
|
/// - `predicate`: The predicate to call to validate the input
|
||||||
/// - Returns: A parser that takes `n` characters from the input
|
/// - Returns: A parser that takes `n` characters from the input
|
||||||
#[cfg(feature = "nom")]
|
|
||||||
pub fn take_where<F, Input>(n: usize, predicate: F) -> impl Fn(Input) -> IResult<Input, Input>
|
pub fn take_where<F, Input>(n: usize, predicate: F) -> impl Fn(Input) -> IResult<Input, Input>
|
||||||
where
|
where
|
||||||
Input: InputTake + InputIter + InputLength + Slice<RangeFrom<usize>>,
|
Input: InputTake + InputIter + InputLength + Slice<RangeFrom<usize>>,
|
||||||
@ -51,7 +47,6 @@ where
|
|||||||
take_while_m_n(n, n, predicate)
|
take_while_m_n(n, n, predicate)
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(feature = "nom")]
|
|
||||||
pub fn exhausted<'a, Parser, R>(inner: Parser) -> impl FnMut(&'a str) -> IResult<&'a str, R>
|
pub fn exhausted<'a, Parser, R>(inner: Parser) -> impl FnMut(&'a str) -> IResult<&'a str, R>
|
||||||
where
|
where
|
||||||
Parser: Fn(&'a str) -> IResult<&'a str, R>,
|
Parser: Fn(&'a str) -> IResult<&'a str, R>,
|
||||||
@ -59,7 +54,7 @@ where
|
|||||||
terminated(inner, eof)
|
terminated(inner, eof)
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(all(test, feature = "nom"))]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use nom::bytes::streaming::take_while;
|
use nom::bytes::streaming::take_while;
|
||||||
|
|
||||||
|
@ -1,10 +1,8 @@
|
|||||||
#[cfg(feature = "nom")]
|
|
||||||
use {
|
use {
|
||||||
crate::traits::IntoResult,
|
crate::traits::IntoResult,
|
||||||
nom::{error::Error, IResult},
|
nom::{error::Error, IResult},
|
||||||
};
|
};
|
||||||
|
|
||||||
#[cfg(feature = "nom")]
|
|
||||||
impl<T, R> IntoResult<T> for IResult<R, T> {
|
impl<T, R> IntoResult<T> for IResult<R, T> {
|
||||||
type Error = nom::Err<Error<R>>;
|
type Error = nom::Err<Error<R>>;
|
||||||
fn into_result(self) -> Result<T, Self::Error> {
|
fn into_result(self) -> Result<T, Self::Error> {
|
||||||
@ -12,7 +10,7 @@ impl<T, R> IntoResult<T> for IResult<R, T> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(all(test, feature = "nom"))]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use nom::character::complete::char as c;
|
use nom::character::complete::char as c;
|
||||||
|
|
||||||
|
@ -1,15 +1,12 @@
|
|||||||
#[cfg(feature = "serde")]
|
|
||||||
use serde::Serialize;
|
use serde::Serialize;
|
||||||
|
|
||||||
#[derive(Serialize)]
|
#[derive(Serialize)]
|
||||||
#[cfg(feature = "serde")]
|
|
||||||
pub struct BaseResponse<T: Serialize> {
|
pub struct BaseResponse<T: Serialize> {
|
||||||
pub version: String,
|
pub version: String,
|
||||||
#[serde(flatten)]
|
#[serde(flatten)]
|
||||||
pub body: T, // T must be a struct (or enum?)
|
pub body: T, // T must be a struct (or enum?)
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(feature = "serde")]
|
|
||||||
impl<T: Serialize> BaseResponse<T> {
|
impl<T: Serialize> BaseResponse<T> {
|
||||||
pub fn new(version: impl Into<String>, body: T) -> Self {
|
pub fn new(version: impl Into<String>, body: T) -> Self {
|
||||||
Self {
|
Self {
|
||||||
@ -19,7 +16,7 @@ impl<T: Serialize> BaseResponse<T> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(all(test, feature = "serde"))]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use super::*;
|
use super::*;
|
||||||
|
|
||||||
|
@ -1,9 +1,7 @@
|
|||||||
#[cfg(feature = "vec")]
|
|
||||||
pub trait Distinct {
|
pub trait Distinct {
|
||||||
fn distinct(&mut self);
|
fn distinct(&mut self);
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(feature = "vec")]
|
|
||||||
impl<T: PartialEq + Clone> Distinct for Vec<T> {
|
impl<T: PartialEq + Clone> Distinct for Vec<T> {
|
||||||
fn distinct(&mut self) {
|
fn distinct(&mut self) {
|
||||||
*self = self.iter().fold(vec![], |mut acc, x| {
|
*self = self.iter().fold(vec![], |mut acc, x| {
|
||||||
@ -15,7 +13,7 @@ impl<T: PartialEq + Clone> Distinct for Vec<T> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(all(test, feature = "vec"))]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use super::*;
|
use super::*;
|
||||||
|
|
||||||
|
@ -1,5 +1,4 @@
|
|||||||
#[macro_export]
|
#[macro_export]
|
||||||
#[cfg(feature = "vec")]
|
|
||||||
macro_rules! map {
|
macro_rules! map {
|
||||||
() => { std::collections::HashMap::new() };
|
() => { std::collections::HashMap::new() };
|
||||||
($($k:expr => $v:expr),* $(,)?) => {
|
($($k:expr => $v:expr),* $(,)?) => {
|
||||||
@ -13,7 +12,7 @@ macro_rules! map {
|
|||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(all(test, feature = "vec"))]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
|
|
||||||
|
@ -1,5 +1,4 @@
|
|||||||
#[macro_export]
|
#[macro_export]
|
||||||
#[cfg(feature = "vec")]
|
|
||||||
macro_rules! matrix {
|
macro_rules! matrix {
|
||||||
($x:expr; $m:expr, $n:expr) => {
|
($x:expr; $m:expr, $n:expr) => {
|
||||||
vec![vec![$x; $n]; $m]
|
vec![vec![$x; $n]; $m]
|
||||||
@ -16,7 +15,7 @@ macro_rules! matrix {
|
|||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(all(test, feature = "vec"))]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
|
@ -1,5 +1,4 @@
|
|||||||
#[macro_export]
|
#[macro_export]
|
||||||
#[cfg(feature = "vec")]
|
|
||||||
macro_rules! set {
|
macro_rules! set {
|
||||||
() => { std::collections::HashSet::new() };
|
() => { std::collections::HashSet::new() };
|
||||||
($($x:expr),* $(,)?) => {
|
($($x:expr),* $(,)?) => {
|
||||||
@ -13,7 +12,7 @@ macro_rules! set {
|
|||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(all(test, feature = "vec"))]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use std::collections::HashSet;
|
use std::collections::HashSet;
|
||||||
|
|
||||||
|
Loading…
x
Reference in New Issue
Block a user