extern crate markup5ever_rcdom as rcdom; extern crate html5ever; use std::{rc::Rc, time::Instant}; use db::{connect, Website}; use html5ever::{parse_document, tendril::TendrilSink, tree_builder::TreeBuilderOpts, ParseOpts}; use rcdom::{Node, RcDom}; use surrealdb::{engine::remote::ws::Client, sql::Thing, Surreal}; use tracing::{debug, info, instrument, trace, trace_span, warn}; use tracing_subscriber::EnvFilter; mod db; #[tokio::main] async fn main() { tracing_subscriber::fmt() .with_env_filter(EnvFilter::from_default_env()) .with_line_number(true) .without_time() .init(); debug!("Starting..."); // Would probably take these in as parameters from a cli let url = "https://oliveratkinson.net/"; let budget = 50; let mut crawled = 0; let db = connect().await.expect("Failed to connect to db, aborting."); // Kick off the whole machine - This Website object doesn't matter, it's just to allow for // get() to work. let span = trace_span!("Pre-Loop"); let pre_loop_span = span.enter(); let mut site = Website::new(&url, false); let dom = get(&mut site, &db).await.expect("Inital page returned None."); crawl_wrapper(&dom, &db, &site, &mut crawled).await; drop(pre_loop_span); let span = trace_span!("Loop"); let span = span.enter(); while crawled < budget { let get_num = if budget - crawled < 100 { budget - crawled } else {100}; let uncrawled = get_uncrawled_links(&db, get_num).await; debug!("Crawling {} pages...", uncrawled.len()); let span = trace_span!("Crawling"); let _ = span.enter(); for mut site in uncrawled { if let Some(dom) = get(&mut site, &db).await { trace!("Pre-walk checkpoint"); crawl_wrapper(&dom, &db, &site, &mut crawled).await; let percent = format!("{:.2}%", (crawled as f32/budget as f32) * 100f32); info!("Crawled {crawled} out of {budget} pages. ({percent})"); } else { warn!("Failed to get {}", site.to_string()); } } } drop(span); info!("Done"); } async fn crawl_wrapper(dom: &Rc, db: &Surreal, site: &Website, count: &mut usize) { let mut buffer = Vec::new(); let now = Instant::now(); walk(&dom, &db, &site, &mut buffer).await; let dif = now.elapsed().as_micros(); trace!("{}", format!("Walked in {:.3}ms", dif as f64/1000.)); site.links_to(buffer, &db).await; *count += 1; } #[instrument(skip_all)] /// A quick helper function for downloading a url async fn get(site: &mut Website, db: &Surreal) -> Option> { trace!("Get: {}", site.to_string()); let now = Instant::now(); if let Ok(response) = reqwest::get(site.to_string()).await { let dif = now.elapsed().as_micros(); trace!("{}", format!("Got page in {:.3}ms", dif as f64/1000.)); let data = response.text().await.unwrap(); let opts = ParseOpts { tree_builder: TreeBuilderOpts { drop_doctype: true, ..Default::default() }, ..Default::default() }; let dom = parse_document(RcDom::default(), opts) .from_utf8() .read_from(&mut data.as_bytes()) .unwrap(); site.set_crawled(); site.store(db).await; trace!("Got: {}", site.to_string()); return Some(dom.document); } trace!("Failed to get: {}", site.to_string()); None } /// Walks the givin site, placing it's findings in the database async fn walk(node: &rcdom::Handle, db: &Surreal , site: &Website, links_to: &mut Vec) { let span = trace_span!("Walk"); let span = span.enter(); match &node.data { rcdom::NodeData::Element { name, attrs, template_contents, mathml_annotation_xml_integration_point } => { for attr in attrs.borrow().clone() { if name.local.to_string() == "a" { if attr.value.starts_with("mailto") { trace!("Is mailto"); // mailto link, lol let _created: Option = db.create("email").content(db::Email { email: attr.value.to_string(), on: site.domain_str().to_owned(), }).await.unwrap(); } else { let mut web = site.clone(); let url = web.mut_url(); // TODO remove #xyz let joined = url.join(&attr.value).unwrap(); *url = joined; let crawled = web.crawled(); *crawled = false; if let Some(id) = web.store(db).await { links_to.push(id); } } } }; }, _ => {}, }; drop(span); for child in node.children.borrow().iter() { Box::pin(walk(child, db, site, links_to)).await; } } /// Returns uncrawled links async fn get_uncrawled_links(db: &Surreal, mut count: usize) -> Vec { if count > 100 { count = 100 } let mut response = db .query("SELECT * FROM website WHERE crawled = false LIMIT $count") .bind(("count", count)) .await .expect("Hard-coded query failed..?"); response.take(0).expect("Returned websites couldn't be parsed") }