pitchfork / scrape_pitchfork.py
mattismegevand's picture
init commit
4d5d3e0
raw
history blame
5.58 kB
import re
import requests
import sqlite3
import pandas as pd
from concurrent.futures import ThreadPoolExecutor
from bs4 import BeautifulSoup, SoupStrainer
from sys import argv
regexes = {
'artist': re.compile(r'SplitScreenContentHeaderArtist-\w*'),
'album': re.compile(r'SplitScreenContentHeaderHed-\w*'),
'year_released': re.compile(r'SplitScreenContentHeaderReleaseYear-\w*'),
'rating': re.compile(r'Rating-\w*'),
'small_text': re.compile(r'SplitScreenContentHeaderDekDown-\w*'),
'review': re.compile(r'body__inner-container'),
'reviewer': re.compile(r'BylineName'),
'genre': re.compile(r'SplitScreenContentHeaderInfoSlice-\w*'),
'label': re.compile(r'SplitScreenContentHeaderInfoSlice-\w*'),
'reviewed': re.compile(r'SplitScreenContentHeaderInfoSlice-\w*'),
'album_art_url': re.compile(r'SplitScreenContentHeaderImage-\w*'),
}
def fetch(args):
''' Fetch a single url and return a dictionary of data from a Pitchfork review '''
url, session = args
response = session.get(url)
if response.status_code == 200:
soup_strainer = SoupStrainer('article', {'data-testid': 'ReviewPageArticle'})
soup = BeautifulSoup(response.content, 'lxml', parse_only=soup_strainer)
if soup.find('article', {'data-testid': 'ReviewPageArticle'}) is None:
with open('not_done.txt', 'a') as f:
f.write(url + '\n')
return None
print('.', end='', flush=True)
result = data_from_soup(soup)
if result is None:
with open('not_done.txt', 'a') as f:
f.write(url + '\n')
return result
else:
with open('errors.txt', 'a') as f:
f.write(url + '\n')
return None
def get_reviews(urls):
''' Return a list of review data dictionaries from the provided urls '''
reviews = []
session = requests.Session()
with ThreadPoolExecutor() as executor:
for result in executor.map(fetch, ((url, session) for url in urls)):
if result: # Check if result is not None
reviews.append(result)
print()
return reviews
def data_from_soup(soup):
''' Return a dictionary of data from a Pitchfork review '''
artist = soup.find('div', {'class': regexes['artist']}).text.strip()
album = soup.find('h1', {'class': regexes['album']}).text.strip()
year_released = soup.find('time', {'class': regexes['year_released']})
if year_released:
year_released = int(year_released.text.strip())
else:
return None
rating = float(soup.find('p', {'class': regexes['rating']}).text.strip())
small_text = soup.find('div', {'class': regexes['small_text']})
small_text = small_text.text.strip() if small_text else 'N/A'
review = "".join(e.text for e in soup.find('div', {'class': regexes['review']}).descendants if e.name == 'p')
reviewer = soup.find('span', {'data-testid': regexes['reviewer']})
reviewer = reviewer.text.strip()[3:] if reviewer else 'N/A'
misc = [e.text for e in soup.find('div', {'class': regexes['genre']}).descendants if e.name == 'li']
misc = {'genre': 'N/A', 'label': 'N/A', 'reviewed': 'N/A'} | {e.split(':')[0].strip().lower(): e.split(':')[1].strip() for e in misc}
album_art_url = soup.find('source', {'media': '(max-width: 767px)'})
album_art_url = album_art_url['srcset'].split(',')[-2].strip() if album_art_url else 'N/A'
return {
'artist': artist, 'album': album, 'year_released': year_released,
'rating': rating, 'small_text': small_text, 'review': review,
'reviewer': reviewer, 'genre': misc['genre'], 'label': misc['label'],
'reviewed': misc['reviewed'], 'album_art_url': album_art_url,
}
def insert_into_db(data, cursor):
''' Insert data into a sqlite3 database '''
for review in data:
artist = review.get('artist')
album = review.get('album')
year_released = review.get('year_released')
rating = review.get('rating')
small_text = review.get('small_text')
review_text = review.get('review') # 'review' is a reserved word in Python
reviewer = review.get('reviewer')
genre = review.get('genre')
label = review.get('label')
reviewed = review.get('reviewed')
album_art_url = review.get('album_art_url')
cursor.execute('SELECT * FROM reviews WHERE artist=? AND album=?', (artist, album))
result = cursor.fetchone()
if result is None:
# Insert new review into database
cursor.execute('''
INSERT INTO reviews (
artist, album, year_released, rating, small_text,
review, reviewer, genre, label, reviewed, album_art_url
) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
''', (
artist, album, year_released, rating, small_text, review_text,
reviewer, genre, label, reviewed, album_art_url
))
def main(start, end):
conn = sqlite3.connect('reviews.db')
c = conn.cursor()
# Create table with all necessary fields
c.execute('''
CREATE TABLE IF NOT EXISTS reviews (
artist TEXT,
album TEXT,
year_released INTEGER,
rating REAL,
small_text TEXT,
review TEXT,
reviewer TEXT,
genre TEXT,
label TEXT,
reviewed TEXT,
album_art_url TEXT
)
''')
# Read URLs from a CSV file into a list
df = pd.read_csv('urls.csv')
urls = df['url'].tolist() # replace 'url' with your actual column name
start, end = max(0, start), min(len(urls), end)
urls = urls[start:end]
print(f'Fetching {len(urls)} reviews')
data = get_reviews(urls)
print(f'Fetching complete. Inserting into database')
insert_into_db(data, c)
print('Done')
conn.commit()
conn.close()
if __name__ == '__main__':
main(int(argv[1]), int(argv[2]))