mirror of
https://github.com/neon-mmd/websurfx.git
synced 2024-11-21 13:38:21 -05:00
updating and improving README.org
This commit is contained in:
parent
e79100069b
commit
ed13a16ec5
@ -1,6 +1,6 @@
|
||||
* Websurfx
|
||||
|
||||
a lightening fast, privacy respecting, secure [[https://en.wikipedia.org/wiki/Metasearch_engine][meta search engine]]. (pronounced as websurface or web-surface //wɛbˈsɜːrfəs//.)
|
||||
A lightening fast, privacy respecting, secure [[https://en.wikipedia.org/wiki/Metasearch_engine][meta search engine]]. (pronounced as websurface or web-surface /wɛbˈsɜːrfəs/.)
|
||||
|
||||
* Preview
|
||||
|
||||
@ -45,7 +45,7 @@ and then open your browser of choice and visit [[http://127.0.0.1:8080]] and the
|
||||
|
||||
** Why Websurfx?
|
||||
|
||||
The main goal of the project is to provide a fast, secure and privacy focused [[https://en.wikipedia.org/wiki/Metasearch_engine][meta search engine]]. Though there are many meta search engine out there but they don't provide gaurantee security of the their search engine which is essential because sometimes privacy is related to security like for example some memory vulnerabilities can leak private or sensitive information which is never good so the project being written in rust gaurantees memory safety and thus eliminating such problems.
|
||||
The main goal of the project is to provide a fast, secure and privacy focused [[https://en.wikipedia.org/wiki/Metasearch_engine][meta search engine]]. Though there are many meta search engine out there but they don't provide gaurantee security of the their search engine which is essential because sometimes privacy is related to security like for example some memory vulnerabilities can leak private or sensitive information which is never good so the project being written in rust gaurantees memory safety and thus eliminating such problems and also many meta search engines lack many features like advanced image search *(which is required by may graphics designers, content creators, etc), proper nsfw blocking (many links are still visible even on strict safe search), etc which *websurfx* aims to provide.
|
||||
|
||||
** Why GPLv3?
|
||||
|
||||
|
@ -1,3 +1,7 @@
|
||||
//! The `duckduckgo` module handles the scraping of results from the duckduckgo search engine
|
||||
//! by querying the upstream duckduckgo search engine with user provided query and with a page
|
||||
//! number if provided.
|
||||
|
||||
use std::collections::HashMap;
|
||||
|
||||
use reqwest::header::USER_AGENT;
|
||||
@ -5,10 +9,22 @@ use scraper::{Html, Selector};
|
||||
|
||||
use crate::search_results_handler::aggregation_models::RawSearchResult;
|
||||
|
||||
// This function scrapes results from the upstream engine duckduckgo and puts all the scraped
|
||||
// results like title, visiting_url (href in html),engine (from which engine it was fetched from)
|
||||
// and description in a RawSearchResult and then adds that to HashMap whose keys are url and
|
||||
// values are RawSearchResult struct and then returns it within a Result enum.
|
||||
/// This function scrapes results from the upstream engine duckduckgo and puts all the scraped
|
||||
/// results like title, visiting_url (href in html),engine (from which engine it was fetched from)
|
||||
/// and description in a RawSearchResult and then adds that to HashMap whose keys are url and
|
||||
/// values are RawSearchResult struct and then returns it within a Result enum.
|
||||
///
|
||||
/// # Arguments
|
||||
///
|
||||
/// * `query` - Takes the user provided query to query to the upstream search engine with.
|
||||
/// * `page` - Takes an Option<u32> as argument which can be either None or a valid page number.
|
||||
/// * `user_agent` - Takes a random user agent string as an argument.
|
||||
///
|
||||
/// # Errors
|
||||
///
|
||||
/// Returns a reqwest error if the user is not connected to the internet or if their is failure to
|
||||
/// reach the above **upstream search engine** page and also returns error if the scraping
|
||||
/// selector fails to initialize"
|
||||
pub async fn results(
|
||||
query: &str,
|
||||
page: Option<u32>,
|
||||
|
@ -1,3 +1,7 @@
|
||||
//! The `searx` module handles the scraping of results from the searx search engine instance
|
||||
//! by querying the upstream searx search engine instance with user provided query and with a page
|
||||
//! number if provided.
|
||||
|
||||
use std::collections::HashMap;
|
||||
|
||||
use reqwest::header::USER_AGENT;
|
||||
@ -5,10 +9,22 @@ use scraper::{Html, Selector};
|
||||
|
||||
use crate::search_results_handler::aggregation_models::RawSearchResult;
|
||||
|
||||
// This function scrapes results from the upstream engine searx instance and puts all the scraped
|
||||
// results like title, visiting_url (href in html),engine (from which engine it was fetched from)
|
||||
// and description in a RawSearchResult and then adds that to HashMap whose keys are url and
|
||||
// values are RawSearchResult struct and then returns it within a Result enum.
|
||||
/// This function scrapes results from the upstream engine duckduckgo and puts all the scraped
|
||||
/// results like title, visiting_url (href in html),engine (from which engine it was fetched from)
|
||||
/// and description in a RawSearchResult and then adds that to HashMap whose keys are url and
|
||||
/// values are RawSearchResult struct and then returns it within a Result enum.
|
||||
///
|
||||
/// # Arguments
|
||||
///
|
||||
/// * `query` - Takes the user provided query to query to the upstream search engine with.
|
||||
/// * `page` - Takes an Option<u32> as argument which can be either None or a valid page number.
|
||||
/// * `user_agent` - Takes a random user agent string as an argument.
|
||||
///
|
||||
/// # Errors
|
||||
///
|
||||
/// Returns a reqwest error if the user is not connected to the internet or if their is failure to
|
||||
/// reach the above **upstream search engine** page and also returns error if the scraping
|
||||
/// selector fails to initialize"
|
||||
pub async fn results(
|
||||
query: &str,
|
||||
page: Option<u32>,
|
||||
|
Loading…
Reference in New Issue
Block a user