package main import ( "bufio" "crypto/md5" "encoding/hex" "encoding/json" "fmt" "io" "io/ioutil" "net/http" "net/url" "os" "regexp" "strings" "sync" "sync/atomic" "time" ) type Record struct { URL string `json:"url"` Text string `json:"text"` TS string `json:"ts"` } var ( visited = make(map[string]bool) visitedMu sync.RWMutex queue = make(chan string, 1000000) stats struct { pages, bytes, errors int64 } client = &http.Client{ Timeout: 5 * time.Second, Transport: &http.Transport{ MaxIdleConns: 1000, MaxIdleConnsPerHost: 50, IdleConnTimeout: 30 * time.Second, }, } linkRe = regexp.MustCompile(`href=["']([^"']+)["']`) scriptRe = regexp.MustCompile(`(?is)]*>.*?|]*>.*?`) tagRe = regexp.MustCompile(`<[^>]+>`) spaceRe = regexp.MustCompile(`\s+`) ) func hash(s string) string { h := md5.Sum([]byte(s)) return hex.EncodeToString(h[:8]) } func extractText(html string) string { text := scriptRe.ReplaceAllString(html, " ") text = tagRe.ReplaceAllString(text, " ") text = spaceRe.ReplaceAllString(text, " ") text = strings.TrimSpace(text) if len(text) > 50000 { text = text[:50000] } return text } func extractLinks(html, baseURL string) []string { base, err := url.Parse(baseURL) if err != nil { return nil } var links []string matches := linkRe.FindAllStringSubmatch(html, 100) for _, m := range matches { if len(m) > 1 { href := m[1] if u, err := url.Parse(href); err == nil { resolved := base.ResolveReference(u) if resolved.Scheme == "http" || resolved.Scheme == "https" { links = append(links, resolved.String()) } } } } return links } func worker(id int, output chan<- Record) { for urlStr := range queue { h := hash(urlStr) visitedMu.RLock() seen := visited[h] visitedMu.RUnlock() if seen { continue } visitedMu.Lock() visited[h] = true visitedMu.Unlock() resp, err := client.Get(urlStr) if err != nil { atomic.AddInt64(&stats.errors, 1) continue } ct := resp.Header.Get("Content-Type") if resp.StatusCode != 200 || !strings.Contains(ct, "text/html") { resp.Body.Close() atomic.AddInt64(&stats.errors, 1) continue } body, _ := ioutil.ReadAll(io.LimitReader(resp.Body, 1<<20)) resp.Body.Close() html := string(body) text := extractText(html) if len(text) > 200 { output <- Record{URL: urlStr, Text: text, TS: time.Now().UTC().Format("2006-01-02T15:04:05")} atomic.AddInt64(&stats.pages, 1) atomic.AddInt64(&stats.bytes, int64(len(text))) } for _, link := range extractLinks(html, urlStr) { select { case queue <- link: default: } } } } func main() { seeds := []string{ "https://en.wikipedia.org/wiki/Main_Page", "https://en.wikipedia.org/wiki/Special:Random", "https://news.ycombinator.com/", "https://www.reddit.com/r/all/", "https://www.reddit.com/r/programming/", "https://stackoverflow.com/questions", "https://medium.com/", "https://dev.to/", } for _, s := range seeds { queue <- s } os.MkdirAll("/workspace/go_crawl", 0755) outPath := fmt.Sprintf("/workspace/go_crawl/crawl_%s.jsonl", time.Now().Format("20060102_150405")) f, _ := os.Create(outPath) defer f.Close() w := bufio.NewWriter(f) defer w.Flush() output := make(chan Record, 10000) go func() { for rec := range output { data, _ := json.Marshal(rec) w.Write(data) w.WriteByte('\n') } }() start := time.Now() go func() { for { time.Sleep(10 * time.Second) elapsed := time.Since(start).Minutes() if elapsed > 0 { p := atomic.LoadInt64(&stats.pages) b := atomic.LoadInt64(&stats.bytes) e := atomic.LoadInt64(&stats.errors) fmt.Printf("[%s] %d pgs | %.0fMB | %.1fMB/min | Q:%d | Err:%d\n", time.Now().Format("15:04:05"), p, float64(b)/1024/1024, float64(b)/1024/1024/elapsed, len(queue), e) } } }() fmt.Printf("=== GO STDLIB CRAWLER ===\nWorkers: 1000\nOutput: %s\n\n", outPath) var wg sync.WaitGroup for i := 0; i < 1000; i++ { wg.Add(1) go func(id int) { defer wg.Done() worker(id, output) }(i) } wg.Wait() close(output) }