OpenTransformer commited on
Commit
a49ab18
·
verified ·
1 Parent(s): e90083c

Upload src/main.rs with huggingface_hub

Browse files
Files changed (1) hide show
  1. src/main.rs +144 -0
src/main.rs ADDED
@@ -0,0 +1,144 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ use std::collections::HashSet;
2
+ use std::fs::OpenOptions;
3
+ use std::io::Write;
4
+ use std::sync::{Arc, Mutex};
5
+ use std::time::{Duration, Instant};
6
+ use tokio::sync::Semaphore;
7
+ use reqwest::Client;
8
+ use scraper::{Html, Selector};
9
+ use url::Url;
10
+
11
+ const MAX_CONCURRENT: usize = 500;
12
+ const OUTPUT_FILE: &str = "/workspace/crawl.jsonl";
13
+
14
+ #[tokio::main]
15
+ async fn main() {
16
+ let client = Client::builder()
17
+ .timeout(Duration::from_secs(10))
18
+ .user_agent("Mozilla/5.0 (compatible; Googlebot/2.1)")
19
+ .build()
20
+ .unwrap();
21
+
22
+ let visited: Arc<Mutex<HashSet<String>>> = Arc::new(Mutex::new(HashSet::new()));
23
+ let queue: Arc<Mutex<Vec<String>>> = Arc::new(Mutex::new(Vec::new()));
24
+ let file = Arc::new(Mutex::new(
25
+ OpenOptions::new().create(true).append(true).open(OUTPUT_FILE).unwrap()
26
+ ));
27
+ let semaphore = Arc::new(Semaphore::new(MAX_CONCURRENT));
28
+
29
+ // Seed URLs
30
+ let seeds = vec![
31
+ "https://en.wikipedia.org", "https://www.bbc.com", "https://arxiv.org",
32
+ "https://www.nature.com", "https://www.theguardian.com", "https://reuters.com",
33
+ "https://www.npr.org", "https://www.sciencedaily.com", "https://arstechnica.com",
34
+ ];
35
+
36
+ {
37
+ let mut q = queue.lock().unwrap();
38
+ for seed in seeds {
39
+ q.push(seed.to_string());
40
+ }
41
+ }
42
+
43
+ let start = Instant::now();
44
+ let pages = Arc::new(Mutex::new(0u64));
45
+ let bytes = Arc::new(Mutex::new(0u64));
46
+
47
+ loop {
48
+ let url = {
49
+ let mut q = queue.lock().unwrap();
50
+ q.pop()
51
+ };
52
+
53
+ let url = match url {
54
+ Some(u) => u,
55
+ None => {
56
+ tokio::time::sleep(Duration::from_millis(100)).await;
57
+ continue;
58
+ }
59
+ };
60
+
61
+ {
62
+ let mut v = visited.lock().unwrap();
63
+ if v.contains(&url) { continue; }
64
+ v.insert(url.clone());
65
+ }
66
+
67
+ let permit = semaphore.clone().acquire_owned().await.unwrap();
68
+ let client = client.clone();
69
+ let queue = queue.clone();
70
+ let file = file.clone();
71
+ let pages = pages.clone();
72
+ let bytes = bytes.clone();
73
+ let visited = visited.clone();
74
+ let start = start.clone();
75
+
76
+ tokio::spawn(async move {
77
+ let _permit = permit;
78
+
79
+ if let Ok(resp) = client.get(&url).send().await {
80
+ if let Ok(text) = resp.text().await {
81
+ let text_len = text.len();
82
+
83
+ // Write to file
84
+ let json = serde_json::json!({
85
+ "url": url,
86
+ "text": text,
87
+ "tokens_est": text_len / 4
88
+ });
89
+
90
+ {
91
+ let mut f = file.lock().unwrap();
92
+ writeln!(f, "{}", json).ok();
93
+ }
94
+
95
+ // Extract links
96
+ let doc = Html::parse_document(&text);
97
+ let selector = Selector::parse("a[href]").unwrap();
98
+ let base = Url::parse(&url).ok();
99
+
100
+ let mut new_urls = Vec::new();
101
+ for elem in doc.select(&selector) {
102
+ if let Some(href) = elem.value().attr("href") {
103
+ if let Some(base) = &base {
104
+ if let Ok(abs) = base.join(href) {
105
+ let abs_str = abs.to_string();
106
+ if abs_str.starts_with("http") {
107
+ new_urls.push(abs_str);
108
+ }
109
+ }
110
+ }
111
+ }
112
+ }
113
+
114
+ {
115
+ let v = visited.lock().unwrap();
116
+ let mut q = queue.lock().unwrap();
117
+ for u in new_urls {
118
+ if !v.contains(&u) && q.len() < 5_000_000 {
119
+ q.push(u);
120
+ }
121
+ }
122
+ }
123
+
124
+ // Update stats
125
+ {
126
+ let mut p = pages.lock().unwrap();
127
+ let mut b = bytes.lock().unwrap();
128
+ *p += 1;
129
+ *b += text_len as u64;
130
+
131
+ if *p % 100 == 0 {
132
+ let elapsed = start.elapsed().as_secs();
133
+ let mb = *b / 1_048_576;
134
+ let qlen = queue.lock().unwrap().len();
135
+ let vlen = visited.lock().unwrap().len();
136
+ eprintln!("📊 {}s | {} pages | {} MB | queue: {} | visited: {}",
137
+ elapsed, p, mb, qlen, vlen);
138
+ }
139
+ }
140
+ }
141
+ }
142
+ });
143
+ }
144
+ }