mirror of
https://github.com/neon-mmd/websurfx.git
synced 2024-11-22 14:08:23 -05:00
✨ feat: implement async multithreading and engine selection code
This commit is contained in:
parent
897ab0807f
commit
0781385393
@ -3,15 +3,20 @@
|
|||||||
|
|
||||||
use std::{collections::HashMap, time::Duration};
|
use std::{collections::HashMap, time::Duration};
|
||||||
|
|
||||||
|
use error_stack::Report;
|
||||||
use rand::Rng;
|
use rand::Rng;
|
||||||
use tokio::join;
|
use tokio::task::JoinHandle;
|
||||||
|
|
||||||
use super::{
|
use super::{
|
||||||
aggregation_models::{RawSearchResult, SearchResult, SearchResults},
|
aggregation_models::{RawSearchResult, SearchResult, SearchResults},
|
||||||
user_agent::random_user_agent,
|
user_agent::random_user_agent,
|
||||||
};
|
};
|
||||||
|
|
||||||
use crate::engines::{duckduckgo, searx};
|
use crate::engines::{
|
||||||
|
duckduckgo,
|
||||||
|
engine_models::{EngineError, SearchEngine},
|
||||||
|
searx,
|
||||||
|
};
|
||||||
|
|
||||||
/// A function that aggregates all the scraped results from the above upstream engines and
|
/// A function that aggregates all the scraped results from the above upstream engines and
|
||||||
/// then removes duplicate results and if two results are found to be from two or more engines
|
/// then removes duplicate results and if two results are found to be from two or more engines
|
||||||
@ -37,10 +42,11 @@ use crate::engines::{duckduckgo, searx};
|
|||||||
/// function in either `searx` or `duckduckgo` or both otherwise returns a `SearchResults struct`
|
/// function in either `searx` or `duckduckgo` or both otherwise returns a `SearchResults struct`
|
||||||
/// containing appropriate values.
|
/// containing appropriate values.
|
||||||
pub async fn aggregate(
|
pub async fn aggregate(
|
||||||
query: &str,
|
query: String,
|
||||||
page: u32,
|
page: u32,
|
||||||
random_delay: bool,
|
random_delay: bool,
|
||||||
debug: bool,
|
debug: bool,
|
||||||
|
upstream_search_engines: Vec<String>,
|
||||||
) -> Result<SearchResults, Box<dyn std::error::Error>> {
|
) -> Result<SearchResults, Box<dyn std::error::Error>> {
|
||||||
let user_agent: String = random_user_agent();
|
let user_agent: String = random_user_agent();
|
||||||
let mut result_map: HashMap<String, RawSearchResult> = HashMap::new();
|
let mut result_map: HashMap<String, RawSearchResult> = HashMap::new();
|
||||||
@ -53,28 +59,59 @@ pub async fn aggregate(
|
|||||||
}
|
}
|
||||||
|
|
||||||
// fetch results from upstream search engines simultaneously/concurrently.
|
// fetch results from upstream search engines simultaneously/concurrently.
|
||||||
let (ddg_map_results, searx_map_results) = join!(
|
let search_engines: Vec<Box<dyn SearchEngine>> = upstream_search_engines
|
||||||
duckduckgo::results(query, page, &user_agent),
|
.iter()
|
||||||
searx::results(query, page, &user_agent)
|
.map(|engine| match engine.to_lowercase().as_str() {
|
||||||
|
"duckduckgo" => Box::new(duckduckgo::DuckDuckGo) as Box<dyn SearchEngine>,
|
||||||
|
"searx " => Box::new(searx::Searx) as Box<dyn SearchEngine>,
|
||||||
|
})
|
||||||
|
.collect();
|
||||||
|
|
||||||
|
let tasks: Vec<JoinHandle<Result<HashMap<String, RawSearchResult>, Report<EngineError>>>> =
|
||||||
|
search_engines
|
||||||
|
.iter()
|
||||||
|
.map(|search_engine| {
|
||||||
|
tokio::spawn(search_engine.results(query.clone(), page, user_agent.clone()))
|
||||||
|
})
|
||||||
|
.collect();
|
||||||
|
|
||||||
|
let mut outputs = Vec::with_capacity(search_engines.len());
|
||||||
|
|
||||||
|
for task in tasks {
|
||||||
|
outputs.push(task.await.ok())
|
||||||
|
}
|
||||||
|
|
||||||
|
let mut initial: bool = true;
|
||||||
|
let mut counter: usize = 0;
|
||||||
|
outputs.iter().for_each(|results| {
|
||||||
|
if initial {
|
||||||
|
match results {
|
||||||
|
Some(result) => {
|
||||||
|
let new_result = result.clone();
|
||||||
|
result_map.extend(new_result.as_ref().unwrap().clone());
|
||||||
|
counter += 1;
|
||||||
|
initial = false
|
||||||
|
}
|
||||||
|
None => {
|
||||||
|
if debug {
|
||||||
|
log::error!(
|
||||||
|
"Error fetching results from {}",
|
||||||
|
upstream_search_engines[counter]
|
||||||
);
|
);
|
||||||
|
};
|
||||||
let ddg_map_results = ddg_map_results.unwrap_or_else(|e| {
|
counter += 1
|
||||||
if debug {
|
|
||||||
log::error!("Error fetching results from DuckDuckGo: {:?}", e);
|
|
||||||
}
|
}
|
||||||
HashMap::new()
|
|
||||||
});
|
|
||||||
|
|
||||||
let searx_map_results = searx_map_results.unwrap_or_else(|e| {
|
|
||||||
if debug {
|
|
||||||
log::error!("Error fetching results from Searx: {:?}", e);
|
|
||||||
}
|
}
|
||||||
HashMap::new()
|
} else {
|
||||||
});
|
match results {
|
||||||
|
Some(result) => {
|
||||||
result_map.extend(ddg_map_results);
|
let new_result = result.clone();
|
||||||
|
new_result
|
||||||
searx_map_results.into_iter().for_each(|(key, value)| {
|
.as_ref()
|
||||||
|
.unwrap()
|
||||||
|
.clone()
|
||||||
|
.into_iter()
|
||||||
|
.for_each(|(key, value)| {
|
||||||
result_map
|
result_map
|
||||||
.entry(key)
|
.entry(key)
|
||||||
.and_modify(|result| {
|
.and_modify(|result| {
|
||||||
@ -89,6 +126,20 @@ pub async fn aggregate(
|
|||||||
)
|
)
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
counter += 1
|
||||||
|
}
|
||||||
|
None => {
|
||||||
|
if debug {
|
||||||
|
log::error!(
|
||||||
|
"Error fetching results from {}",
|
||||||
|
upstream_search_engines[counter]
|
||||||
|
);
|
||||||
|
};
|
||||||
|
counter += 1
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
Ok(SearchResults::new(
|
Ok(SearchResults::new(
|
||||||
result_map
|
result_map
|
||||||
|
Loading…
Reference in New Issue
Block a user