crabbysearch/src/engines/duckduckgo.rs

118 lines
3.9 KiB
Rust
Raw Normal View History

2023-04-27 16:06:59 +03:00
//! The `duckduckgo` module handles the scraping of results from the duckduckgo search engine
//! by querying the upstream duckduckgo search engine with user provided query and with a page
//! number if provided.
2023-07-14 17:16:13 +03:00
use std::collections::HashMap;
2023-04-22 14:35:07 +03:00
use reqwest::header::HeaderMap;
2023-09-24 13:54:08 +02:00
use scraper::Html;
2023-04-22 14:35:07 +03:00
use crate::models::aggregation_models::SearchResult;
2023-04-22 14:35:07 +03:00
use crate::models::engine_models::{EngineError, SearchEngine};
use error_stack::{Report, Result, ResultExt};
2023-09-24 13:54:08 +02:00
use super::search_result_parser::SearchResultParser;
2023-07-15 13:36:46 +03:00
/// A new DuckDuckGo engine type defined in-order to implement the `SearchEngine` trait which allows to
/// reduce code duplication as well as allows to create vector of different search engines easily.
2023-09-24 13:54:08 +02:00
pub struct DuckDuckGo {
parser: SearchResultParser,
}
impl DuckDuckGo {
pub fn new() -> Result<Self, EngineError> {
Ok(Self {
parser: SearchResultParser::new(
".no-results",
".result",
".result__a",
".result__url",
".result__snippet",
)?,
})
}
}
#[async_trait::async_trait]
impl SearchEngine for DuckDuckGo {
2023-09-03 19:34:22 +03:00
async fn results(
&self,
query: &str,
page: u32,
user_agent: &str,
request_timeout: u8,
_safe_search: u8,
) -> Result<HashMap<String, SearchResult>, EngineError> {
// Page number can be missing or empty string and so appropriate handling is required
// so that upstream server recieves valid page number.
let url: String = match page {
1 | 0 => {
format!("https://html.duckduckgo.com/html/?q={query}&s=&dc=&v=1&o=json&api=/d.js")
}
_ => {
format!(
"https://duckduckgo.com/html/?q={}&s={}&dc={}&v=1&o=json&api=/d.js",
query,
(page / 2 + (page % 2)) * 30,
(page / 2 + (page % 2)) * 30 + 1
)
}
};
2023-04-22 14:35:07 +03:00
// initializing HeaderMap and adding appropriate headers.
let header_map = HeaderMap::try_from(&HashMap::from([
("USER_AGENT".to_string(), user_agent.to_string()),
("REFERER".to_string(), "https://google.com/".to_string()),
(
"CONTENT_TYPE".to_string(),
"application/x-www-form-urlencoded".to_string(),
),
("COOKIE".to_string(), "kl=wt-wt".to_string()),
]))
.change_context(EngineError::UnexpectedError)?;
let document: Html = Html::parse_document(
&DuckDuckGo::fetch_html_from_upstream(self, &url, header_map, request_timeout).await?,
);
2023-09-24 13:54:08 +02:00
if document.select(&self.parser.no_result).next().is_some() {
return Err(Report::new(EngineError::EmptyResultSet));
}
2023-04-22 14:35:07 +03:00
// scrape all the results from the html
Ok(document
2023-09-24 13:54:08 +02:00
.select(&self.parser.results)
.map(|result| {
SearchResult::new(
result
2023-09-24 13:54:08 +02:00
.select(&self.parser.result_title)
.next()
.unwrap()
.inner_html()
.trim(),
format!(
"https://{}",
result
2023-09-24 13:54:08 +02:00
.select(&self.parser.result_url)
.next()
.unwrap()
.inner_html()
.trim()
)
.as_str(),
result
2023-09-24 13:54:08 +02:00
.select(&self.parser.result_desc)
.next()
.unwrap()
.inner_html()
.trim(),
&["duckduckgo"],
)
})
.map(|search_result| (search_result.url.clone(), search_result))
.collect())
}
2023-04-22 14:35:07 +03:00
}