0
0
mirror of https://github.com/neon-mmd/websurfx.git synced 2024-11-22 14:08:23 -05:00
This commit is contained in:
Łukasz Mariański 2024-09-08 03:26:49 +00:00 committed by GitHub
commit 9fa0c0a790
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
5 changed files with 44 additions and 8 deletions

13
Cargo.lock generated
View File

@ -3178,6 +3178,7 @@ dependencies = [
"sync_wrapper",
"tokio 1.38.0",
"tokio-rustls",
"tokio-socks",
"tokio-util",
"tower-service",
"url 2.5.0",
@ -4032,6 +4033,18 @@ dependencies = [
"tokio 1.38.0",
]
[[package]]
name = "tokio-socks"
version = "0.5.1"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "51165dfa029d2a65969413a6cc96f354b86b464498702f174a4efa13608fd8c0"
dependencies = [
"either",
"futures-util",
"thiserror",
"tokio 1.36.0",
]
[[package]]
name = "tokio-sync"
version = "0.1.8"

View File

@ -17,7 +17,8 @@ reqwest = { version = "0.12.5", default-features = false, features = [
"rustls-tls",
"brotli",
"gzip",
"http2"
"socks",
"http2",
] }
tokio = { version = "1.32.0", features = [
"rt-multi-thread",

View File

@ -6,6 +6,7 @@ use crate::handler::{file_path, FileType};
use crate::models::parser_models::{AggregatorConfig, RateLimiter, Style};
use log::LevelFilter;
use mlua::Lua;
use reqwest::Proxy;
use std::{collections::HashMap, fs, thread::available_parallelism};
/// A named struct which stores the parsed config file options.
@ -48,6 +49,9 @@ pub struct Config {
pub tcp_connection_keep_alive: u8,
/// It stores the pool idle connection timeout in seconds.
pub pool_idle_connection_timeout: u8,
/// Url of the proxy to use for outgoing requests.
pub proxy: Option<Proxy>,
}
impl Config {
@ -118,6 +122,15 @@ impl Config {
_ => parsed_cet,
};
let proxy_str = globals.get::<_, String>("proxy")?;
let proxy = match Proxy::all(proxy_str) {
Ok(proxy) => Some(proxy),
Err(_) => {
log::error!("Invalid proxy url, defaulting to no proxy.");
None
}
};
Ok(Config {
port: globals.get::<_, u16>("port")?,
binding_ip: globals.get::<_, String>("binding_ip")?,
@ -148,6 +161,7 @@ impl Config {
safe_search,
#[cfg(any(feature = "redis-cache", feature = "memory-cache"))]
cache_expiry_time,
proxy,
})
}
}

View File

@ -76,7 +76,7 @@ pub async fn aggregate(
safe_search: u8,
) -> Result<SearchResults, Box<dyn std::error::Error>> {
let client = CLIENT.get_or_init(|| {
ClientBuilder::new()
let mut cb = ClientBuilder::new()
.timeout(Duration::from_secs(config.request_timeout as u64)) // Add timeout to request to avoid DDOSing the server
.pool_idle_timeout(Duration::from_secs(
config.pool_idle_connection_timeout as u64,
@ -86,9 +86,13 @@ pub async fn aggregate(
.https_only(true)
.gzip(true)
.brotli(true)
.http2_adaptive_window(config.adaptive_window)
.build()
.unwrap()
.http2_adaptive_window(config.adaptive_window);
if config.proxy.is_some() {
cb = cb.proxy(config.proxy.clone().unwrap());
}
cb.build().unwrap()
});
let user_agent: &str = random_user_agent();
@ -247,6 +251,7 @@ pub async fn filter_with_lists(
Ok(())
}
/// Sorts SearchResults by relevance score.
/// <br> sort_unstable is used as its faster,stability is not an issue on our side.
/// For reasons why, check out [`this`](https://rust-lang.github.io/rfcs/1884-unstable-sort.html)
@ -262,6 +267,7 @@ fn sort_search_results(results: &mut [SearchResult]) {
.unwrap_or(Ordering::Less)
})
}
#[cfg(test)]
mod tests {
use super::*;
@ -291,7 +297,7 @@ mod tests {
url: "https://www.rust-lang.org/".to_owned(),
description: "A systems programming language that runs blazingly fast, prevents segfaults, and guarantees thread safety.".to_owned(),
engine: smallvec!["Google".to_owned(), "DuckDuckGo".to_owned()],
relevance_score:0.0
relevance_score: 0.0,
}, )
);
@ -342,7 +348,7 @@ mod tests {
url: "https://www.rust-lang.org/".to_owned(),
description: "A systems programming language that runs blazingly fast, prevents segfaults, and guarantees thread safety.".to_owned(),
engine: smallvec!["Google".to_owned(), "DuckDuckGo".to_owned()],
relevance_score:0.0
relevance_score: 0.0,
},
));

View File

@ -73,3 +73,5 @@ upstream_search_engines = {
Mojeek = false,
Bing = false,
} -- select the upstream search engines from which the results should be fetched.
proxy = "" -- Proxy to send outgoing requests through. Set to empty string to disable.