0
0
mirror of https://github.com/neon-mmd/websurfx.git synced 2024-11-24 06:58:22 -05:00

fix spelling

This commit is contained in:
Milim 2023-06-29 19:18:42 +02:00
parent e2907edf6e
commit dbe5b533d6
6 changed files with 8 additions and 8 deletions

View File

@ -16,7 +16,7 @@
## Author's checklist ## Author's checklist
<!-- additional notes for reviewiers --> <!-- additional notes for reviewers -->
## Related issues ## Related issues

View File

@ -59,7 +59,7 @@
- **Community** - **Community**
- [📊 System Requirements](#system-requirements-) - [📊 System Requirements](#system-requirements-)
- [🗨️ FAQ (Frequently Asked Questions)](#faq-frequently-asked-questions-) - [🗨️ FAQ (Frequently Asked Questions)](#faq-frequently-asked-questions-)
- [📣 More Contributers Wanted](#more-contributers-wanted-) - [📣 More Contributors Wanted](#more-contributers-wanted-)
- [💖 Supporting Websurfx](#supporting-websurfx-) - [💖 Supporting Websurfx](#supporting-websurfx-)
- [📘 Documentation](#documentation-) - [📘 Documentation](#documentation-)
- [🛣️ Roadmap](#roadmap-) - [🛣️ Roadmap](#roadmap-)
@ -165,7 +165,7 @@ Websurfx is based on Rust due to its memory safety features, which prevents vuln
**[⬆️ Back to Top](#--)** **[⬆️ Back to Top](#--)**
# More Contributers Wanted 📣 # More Contributors Wanted 📣
We are looking for more willing contributors to help grow this project. For more information on how you can contribute, check out the [project board](https://github.com/neon-mmd/websurfx/projects?query=is%3Aopen) and the [CONTRIBUTING.md](CONTRIBUTING.md) file for guidelines and rules for making contributions. We are looking for more willing contributors to help grow this project. For more information on how you can contribute, check out the [project board](https://github.com/neon-mmd/websurfx/projects?query=is%3Aopen) and the [CONTRIBUTING.md](CONTRIBUTING.md) file for guidelines and rules for making contributions.

View File

@ -36,7 +36,7 @@ pub async fn results(
user_agent: &str, user_agent: &str,
) -> Result<HashMap<String, RawSearchResult>, EngineError> { ) -> Result<HashMap<String, RawSearchResult>, EngineError> {
// Page number can be missing or empty string and so appropriate handling is required // Page number can be missing or empty string and so appropriate handling is required
// so that upstream server recieves valid page number. // so that upstream server receives valid page number.
let url: String = match page { let url: String = match page {
1 => { 1 => {
format!("https://html.duckduckgo.com/html/?q={query}&s=&dc=&v=1&o=json&api=/d.js") format!("https://html.duckduckgo.com/html/?q={query}&s=&dc=&v=1&o=json&api=/d.js")
@ -86,7 +86,7 @@ pub async fn results(
let results: String = reqwest::Client::new() let results: String = reqwest::Client::new()
.get(url) .get(url)
.timeout(Duration::from_secs(5)) .timeout(Duration::from_secs(5))
.headers(header_map) // add spoofed headers to emulate human behaviour .headers(header_map) // add spoofed headers to emulate human behavior
.send() .send()
.await .await
.into_report() .into_report()

View File

@ -13,7 +13,7 @@ use std::fmt;
/// search engines. /// search engines.
/// * `UnexpectedError` - This variant handles all the errors which are unexpected or occur rarely /// * `UnexpectedError` - This variant handles all the errors which are unexpected or occur rarely
/// and are errors mostly related to failure in initialization of HeaderMap, Selector errors and /// and are errors mostly related to failure in initialization of HeaderMap, Selector errors and
/// all other errors occuring within the code handling the `upstream search engines`. /// all other errors occurring within the code handling the `upstream search engines`.
#[derive(Debug)] #[derive(Debug)]
pub enum EngineError { pub enum EngineError {
EmptyResultSet, EmptyResultSet,

View File

@ -17,7 +17,7 @@ use crate::engines::{duckduckgo, searx};
/// then removes duplicate results and if two results are found to be from two or more engines /// then removes duplicate results and if two results are found to be from two or more engines
/// then puts their names together to show the results are fetched from these upstream engines /// then puts their names together to show the results are fetched from these upstream engines
/// and then removes all data from the HashMap and puts into a struct of all results aggregated /// and then removes all data from the HashMap and puts into a struct of all results aggregated
/// into a vector and also adds the query used into the struct this is neccessory because /// into a vector and also adds the query used into the struct this is necessary because
/// otherwise the search bar in search remains empty if searched from the query url /// otherwise the search bar in search remains empty if searched from the query url
/// ///
/// # Example: /// # Example: