| import re |
| import requests |
| import sqlite3 |
| import pandas as pd |
|
|
| from concurrent.futures import ThreadPoolExecutor |
| from bs4 import BeautifulSoup, SoupStrainer |
| from sys import argv |
|
|
| regexes = { |
| 'artist': re.compile(r'SplitScreenContentHeaderArtist-\w*'), |
| 'album': re.compile(r'SplitScreenContentHeaderHed-\w*'), |
| 'year_released': re.compile(r'SplitScreenContentHeaderReleaseYear-\w*'), |
| 'rating': re.compile(r'Rating-\w*'), |
| 'small_text': re.compile(r'SplitScreenContentHeaderDekDown-\w*'), |
| 'review': re.compile(r'body__inner-container'), |
| 'reviewer': re.compile(r'BylineName'), |
| 'genre': re.compile(r'SplitScreenContentHeaderInfoSlice-\w*'), |
| 'label': re.compile(r'SplitScreenContentHeaderInfoSlice-\w*'), |
| 'reviewed': re.compile(r'SplitScreenContentHeaderInfoSlice-\w*'), |
| 'album_art_url': re.compile(r'SplitScreenContentHeaderImage-\w*'), |
| } |
|
|
| def fetch(args): |
| ''' Fetch a single url and return a dictionary of data from a Pitchfork review ''' |
| url, session = args |
| response = session.get(url) |
| if response.status_code == 200: |
| soup_strainer = SoupStrainer('article', {'data-testid': 'ReviewPageArticle'}) |
| soup = BeautifulSoup(response.content, 'lxml', parse_only=soup_strainer) |
| if soup.find('article', {'data-testid': 'ReviewPageArticle'}) is None: |
| with open('not_done.txt', 'a') as f: |
| f.write(url + '\n') |
| return None |
| print('.', end='', flush=True) |
| result = data_from_soup(soup) |
| if result is None: |
| with open('not_done.txt', 'a') as f: |
| f.write(url + '\n') |
| return result |
| else: |
| with open('errors.txt', 'a') as f: |
| f.write(url + '\n') |
| return None |
|
|
| def get_reviews(urls): |
| ''' Return a list of review data dictionaries from the provided urls ''' |
| reviews = [] |
| session = requests.Session() |
| with ThreadPoolExecutor() as executor: |
| for result in executor.map(fetch, ((url, session) for url in urls)): |
| if result: |
| reviews.append(result) |
| print() |
| return reviews |
|
|
| def data_from_soup(soup): |
| ''' Return a dictionary of data from a Pitchfork review ''' |
| artist = soup.find('div', {'class': regexes['artist']}).text.strip() |
| album = soup.find('h1', {'class': regexes['album']}).text.strip() |
| year_released = soup.find('time', {'class': regexes['year_released']}) |
| if year_released: |
| year_released = int(year_released.text.strip()) |
| else: |
| return None |
| rating = float(soup.find('p', {'class': regexes['rating']}).text.strip()) |
| small_text = soup.find('div', {'class': regexes['small_text']}) |
| small_text = small_text.text.strip() if small_text else 'N/A' |
| review = "".join(e.text for e in soup.find('div', {'class': regexes['review']}).descendants if e.name == 'p') |
| reviewer = soup.find('span', {'data-testid': regexes['reviewer']}) |
| reviewer = reviewer.text.strip()[3:] if reviewer else 'N/A' |
| misc = [e.text for e in soup.find('div', {'class': regexes['genre']}).descendants if e.name == 'li'] |
| misc = {'genre': 'N/A', 'label': 'N/A', 'reviewed': 'N/A'} | {e.split(':')[0].strip().lower(): e.split(':')[1].strip() for e in misc} |
| album_art_url = soup.find('source', {'media': '(max-width: 767px)'}) |
| album_art_url = album_art_url['srcset'].split(',')[-2].strip() if album_art_url else 'N/A' |
| return { |
| 'artist': artist, 'album': album, 'year_released': year_released, |
| 'rating': rating, 'small_text': small_text, 'review': review, |
| 'reviewer': reviewer, 'genre': misc['genre'], 'label': misc['label'], |
| 'reviewed': misc['reviewed'], 'album_art_url': album_art_url, |
| } |
|
|
| def insert_into_db(data, cursor): |
| ''' Insert data into a sqlite3 database ''' |
| for review in data: |
| artist = review.get('artist') |
| album = review.get('album') |
| year_released = review.get('year_released') |
| rating = review.get('rating') |
| small_text = review.get('small_text') |
| review_text = review.get('review') |
| reviewer = review.get('reviewer') |
| genre = review.get('genre') |
| label = review.get('label') |
| reviewed = review.get('reviewed') |
| album_art_url = review.get('album_art_url') |
|
|
| cursor.execute('SELECT * FROM reviews WHERE artist=? AND album=?', (artist, album)) |
| result = cursor.fetchone() |
| if result is None: |
| |
| cursor.execute(''' |
| INSERT INTO reviews ( |
| artist, album, year_released, rating, small_text, |
| review, reviewer, genre, label, reviewed, album_art_url |
| ) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?) |
| ''', ( |
| artist, album, year_released, rating, small_text, review_text, |
| reviewer, genre, label, reviewed, album_art_url |
| )) |
|
|
| def main(start, end): |
| conn = sqlite3.connect('reviews.db') |
| c = conn.cursor() |
|
|
| |
| c.execute(''' |
| CREATE TABLE IF NOT EXISTS reviews ( |
| artist TEXT, |
| album TEXT, |
| year_released INTEGER, |
| rating REAL, |
| small_text TEXT, |
| review TEXT, |
| reviewer TEXT, |
| genre TEXT, |
| label TEXT, |
| reviewed TEXT, |
| album_art_url TEXT |
| ) |
| ''') |
|
|
| |
| df = pd.read_csv('urls.csv') |
| urls = df['url'].tolist() |
| start, end = max(0, start), min(len(urls), end) |
| urls = urls[start:end] |
|
|
| print(f'Fetching {len(urls)} reviews') |
| data = get_reviews(urls) |
| print(f'Fetching complete. Inserting into database') |
| insert_into_db(data, c) |
| print('Done') |
|
|
| conn.commit() |
| conn.close() |
|
|
| if __name__ == '__main__': |
| main(int(argv[1]), int(argv[2])) |
|
|