Compare commits
	
		
			3 Commits
		
	
	
		
			6fc71c7a78
			...
			add6f00ed6
		
	
	| Author | SHA1 | Date | |
|---|---|---|---|
| add6f00ed6 | |||
| 4a433a1a77 | |||
| 03cbcd9ae0 | 
@@ -3,8 +3,9 @@ surreal_url = "localhost:8000"
 | 
				
			|||||||
surreal_username = "root"
 | 
					surreal_username = "root"
 | 
				
			||||||
surreal_password = "root"
 | 
					surreal_password = "root"
 | 
				
			||||||
surreal_ns = "test"
 | 
					surreal_ns = "test"
 | 
				
			||||||
surreal_db = "v1.19.2"
 | 
					surreal_db = "v1.19.5"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
# Crawler config
 | 
					# Crawler config
 | 
				
			||||||
crawl_filter = "en.wikipedia.com" 
 | 
					crawl_filter = "en.wikipedia.org" 
 | 
				
			||||||
budget = 1000
 | 
					start_url = "https://en.wikipedia.org"
 | 
				
			||||||
 | 
					budget = 100
 | 
				
			||||||
 
 | 
				
			|||||||
@@ -4,11 +4,12 @@ Crawls sites saving all the found links to a surrealdb database. It then proceed
 | 
				
			|||||||
 | 
					
 | 
				
			||||||
### TODO
 | 
					### TODO
 | 
				
			||||||
 | 
					
 | 
				
			||||||
- [ ] Domain filtering - prevent the crawler from going on alternate versions of wikipedia.
 | 
					- [x] Domain filtering - prevent the crawler from going on alternate versions of wikipedia.
 | 
				
			||||||
- [ ] Conditionally save content - based on filename or file contents
 | 
					- [ ] Conditionally save content - based on filename or file contents
 | 
				
			||||||
- [x] GUI / TUI ? - Graphana
 | 
					- [x] GUI / TUI ? - Graphana
 | 
				
			||||||
- [x] Better asynchronous getting of the sites. Currently it all happens serially.
 | 
					- [x] Better asynchronous getting of the sites. Currently it all happens serially.
 | 
				
			||||||
- [ ] Allow for storing asynchronously
 | 
					- [x] Allow for storing asynchronously - dropping the "links to" logic fixes this need
 | 
				
			||||||
 | 
					- [x] Control crawler via config file (no recompliation needed)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
3/17/25: Took >1hr to crawl 100 pages
 | 
					3/17/25: Took >1hr to crawl 100 pages
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 
 | 
				
			|||||||
@@ -66,4 +66,3 @@ volumes:
 | 
				
			|||||||
  grafana_storage:
 | 
					  grafana_storage:
 | 
				
			||||||
  alloy_storage:
 | 
					  alloy_storage:
 | 
				
			||||||
  surrealdb_storage:
 | 
					  surrealdb_storage:
 | 
				
			||||||
  minio_storage:
 | 
					 | 
				
			||||||
 
 | 
				
			|||||||
@@ -49,6 +49,7 @@ impl Website {
 | 
				
			|||||||
 | 
					
 | 
				
			||||||
    // Insert ever item in the vec into surreal, crawled state will be preserved as TRUE
 | 
					    // Insert ever item in the vec into surreal, crawled state will be preserved as TRUE
 | 
				
			||||||
    // if already in the database as such or incoming data is TRUE.
 | 
					    // if already in the database as such or incoming data is TRUE.
 | 
				
			||||||
 | 
					    #[instrument(skip(db))]
 | 
				
			||||||
    pub async fn store_all(all: Vec<Self>, db: &Surreal<Client>) -> Vec<Thing> {
 | 
					    pub async fn store_all(all: Vec<Self>, db: &Surreal<Client>) -> Vec<Thing> {
 | 
				
			||||||
        counter!(STORE).increment(1);
 | 
					        counter!(STORE).increment(1);
 | 
				
			||||||
        let mut things = Vec::with_capacity(all.len());
 | 
					        let mut things = Vec::with_capacity(all.len());
 | 
				
			||||||
 
 | 
				
			|||||||
@@ -59,6 +59,13 @@ fn valid_file_extension(take: &&OsStr) -> bool {
 | 
				
			|||||||
                "pdf" => true,
 | 
					                "pdf" => true,
 | 
				
			||||||
                "json" => true,
 | 
					                "json" => true,
 | 
				
			||||||
                "xml" => true,
 | 
					                "xml" => true,
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					                // IGNORE
 | 
				
			||||||
 | 
					                // TODO Should this be a list of all domains?
 | 
				
			||||||
 | 
					                "org" => false,
 | 
				
			||||||
 | 
					                "com" => false,
 | 
				
			||||||
 | 
					                "net" => false,
 | 
				
			||||||
 | 
					
 | 
				
			||||||
                _ => {
 | 
					                _ => {
 | 
				
			||||||
                    warn!("Might be forgetting a file extension: {s}");
 | 
					                    warn!("Might be forgetting a file extension: {s}");
 | 
				
			||||||
                    false
 | 
					                    false
 | 
				
			||||||
 
 | 
				
			|||||||
							
								
								
									
										16
									
								
								src/main.rs
									
									
									
									
									
								
							
							
						
						
									
										16
									
								
								src/main.rs
									
									
									
									
									
								
							@@ -33,11 +33,14 @@ struct Config {
 | 
				
			|||||||
    surreal_password: String,
 | 
					    surreal_password: String,
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    crawl_filter: String,
 | 
					    crawl_filter: String,
 | 
				
			||||||
 | 
					    start_url: String,
 | 
				
			||||||
    budget: usize,
 | 
					    budget: usize,
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#[tokio::main]
 | 
					#[tokio::main]
 | 
				
			||||||
async fn main() {
 | 
					async fn main() {
 | 
				
			||||||
 | 
					    println!("Logs and metrics are provided to the Grafana dashboard");
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    let writer = std::fs::OpenOptions::new()
 | 
					    let writer = std::fs::OpenOptions::new()
 | 
				
			||||||
        .append(true)
 | 
					        .append(true)
 | 
				
			||||||
        .create(true)
 | 
					        .create(true)
 | 
				
			||||||
@@ -70,8 +73,7 @@ async fn main() {
 | 
				
			|||||||
        .expect("failed to install recorder/exporter");
 | 
					        .expect("failed to install recorder/exporter");
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    info!("Starting...");
 | 
					    info!("Starting...");
 | 
				
			||||||
    // Would probably take these in as parameters from a cli
 | 
					
 | 
				
			||||||
    let starting_url = "https://en.wikipedia.org/";
 | 
					 | 
				
			||||||
    // When getting uncrawled pages, name must contain this variable. "" will effectively get ignored.
 | 
					    // When getting uncrawled pages, name must contain this variable. "" will effectively get ignored.
 | 
				
			||||||
    // let crawl_filter = "en.wikipedia.org/";
 | 
					    // let crawl_filter = "en.wikipedia.org/";
 | 
				
			||||||
    // let budget = 50;
 | 
					    // let budget = 50;
 | 
				
			||||||
@@ -82,6 +84,7 @@ async fn main() {
 | 
				
			|||||||
    let _ = file.read_to_string(&mut buf);
 | 
					    let _ = file.read_to_string(&mut buf);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    let config: Config = toml::from_str(&buf).expect("Failed to parse Crawler.toml");
 | 
					    let config: Config = toml::from_str(&buf).expect("Failed to parse Crawler.toml");
 | 
				
			||||||
 | 
					    let starting_url = &config.start_url;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    let db = connect(&config)
 | 
					    let db = connect(&config)
 | 
				
			||||||
        .await
 | 
					        .await
 | 
				
			||||||
@@ -138,6 +141,15 @@ async fn main() {
 | 
				
			|||||||
    }
 | 
					    }
 | 
				
			||||||
    drop(span);
 | 
					    drop(span);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    if let Ok(mut ok) = db.query("count(select id from website where crawled = true)").await {
 | 
				
			||||||
 | 
					        let res = ok.take::<Option<usize>>(0);
 | 
				
			||||||
 | 
					        if let Ok(i) = res {
 | 
				
			||||||
 | 
					            if let Some(n) = i {
 | 
				
			||||||
 | 
					                info!("Total crawled pages now equals {n}");
 | 
				
			||||||
 | 
					            }
 | 
				
			||||||
 | 
					        }
 | 
				
			||||||
 | 
					    }
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    info!("Done");
 | 
					    info!("Done");
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 
 | 
				
			|||||||
		Reference in New Issue
	
	Block a user