Datasets:
File size: 5,576 Bytes
4d5d3e0 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 |
import re
import requests
import sqlite3
import pandas as pd
from concurrent.futures import ThreadPoolExecutor
from bs4 import BeautifulSoup, SoupStrainer
from sys import argv
regexes = {
'artist': re.compile(r'SplitScreenContentHeaderArtist-\w*'),
'album': re.compile(r'SplitScreenContentHeaderHed-\w*'),
'year_released': re.compile(r'SplitScreenContentHeaderReleaseYear-\w*'),
'rating': re.compile(r'Rating-\w*'),
'small_text': re.compile(r'SplitScreenContentHeaderDekDown-\w*'),
'review': re.compile(r'body__inner-container'),
'reviewer': re.compile(r'BylineName'),
'genre': re.compile(r'SplitScreenContentHeaderInfoSlice-\w*'),
'label': re.compile(r'SplitScreenContentHeaderInfoSlice-\w*'),
'reviewed': re.compile(r'SplitScreenContentHeaderInfoSlice-\w*'),
'album_art_url': re.compile(r'SplitScreenContentHeaderImage-\w*'),
}
def fetch(args):
''' Fetch a single url and return a dictionary of data from a Pitchfork review '''
url, session = args
response = session.get(url)
if response.status_code == 200:
soup_strainer = SoupStrainer('article', {'data-testid': 'ReviewPageArticle'})
soup = BeautifulSoup(response.content, 'lxml', parse_only=soup_strainer)
if soup.find('article', {'data-testid': 'ReviewPageArticle'}) is None:
with open('not_done.txt', 'a') as f:
f.write(url + '\n')
return None
print('.', end='', flush=True)
result = data_from_soup(soup)
if result is None:
with open('not_done.txt', 'a') as f:
f.write(url + '\n')
return result
else:
with open('errors.txt', 'a') as f:
f.write(url + '\n')
return None
def get_reviews(urls):
''' Return a list of review data dictionaries from the provided urls '''
reviews = []
session = requests.Session()
with ThreadPoolExecutor() as executor:
for result in executor.map(fetch, ((url, session) for url in urls)):
if result: # Check if result is not None
reviews.append(result)
print()
return reviews
def data_from_soup(soup):
''' Return a dictionary of data from a Pitchfork review '''
artist = soup.find('div', {'class': regexes['artist']}).text.strip()
album = soup.find('h1', {'class': regexes['album']}).text.strip()
year_released = soup.find('time', {'class': regexes['year_released']})
if year_released:
year_released = int(year_released.text.strip())
else:
return None
rating = float(soup.find('p', {'class': regexes['rating']}).text.strip())
small_text = soup.find('div', {'class': regexes['small_text']})
small_text = small_text.text.strip() if small_text else 'N/A'
review = "".join(e.text for e in soup.find('div', {'class': regexes['review']}).descendants if e.name == 'p')
reviewer = soup.find('span', {'data-testid': regexes['reviewer']})
reviewer = reviewer.text.strip()[3:] if reviewer else 'N/A'
misc = [e.text for e in soup.find('div', {'class': regexes['genre']}).descendants if e.name == 'li']
misc = {'genre': 'N/A', 'label': 'N/A', 'reviewed': 'N/A'} | {e.split(':')[0].strip().lower(): e.split(':')[1].strip() for e in misc}
album_art_url = soup.find('source', {'media': '(max-width: 767px)'})
album_art_url = album_art_url['srcset'].split(',')[-2].strip() if album_art_url else 'N/A'
return {
'artist': artist, 'album': album, 'year_released': year_released,
'rating': rating, 'small_text': small_text, 'review': review,
'reviewer': reviewer, 'genre': misc['genre'], 'label': misc['label'],
'reviewed': misc['reviewed'], 'album_art_url': album_art_url,
}
def insert_into_db(data, cursor):
''' Insert data into a sqlite3 database '''
for review in data:
artist = review.get('artist')
album = review.get('album')
year_released = review.get('year_released')
rating = review.get('rating')
small_text = review.get('small_text')
review_text = review.get('review') # 'review' is a reserved word in Python
reviewer = review.get('reviewer')
genre = review.get('genre')
label = review.get('label')
reviewed = review.get('reviewed')
album_art_url = review.get('album_art_url')
cursor.execute('SELECT * FROM reviews WHERE artist=? AND album=?', (artist, album))
result = cursor.fetchone()
if result is None:
# Insert new review into database
cursor.execute('''
INSERT INTO reviews (
artist, album, year_released, rating, small_text,
review, reviewer, genre, label, reviewed, album_art_url
) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
''', (
artist, album, year_released, rating, small_text, review_text,
reviewer, genre, label, reviewed, album_art_url
))
def main(start, end):
conn = sqlite3.connect('reviews.db')
c = conn.cursor()
# Create table with all necessary fields
c.execute('''
CREATE TABLE IF NOT EXISTS reviews (
artist TEXT,
album TEXT,
year_released INTEGER,
rating REAL,
small_text TEXT,
review TEXT,
reviewer TEXT,
genre TEXT,
label TEXT,
reviewed TEXT,
album_art_url TEXT
)
''')
# Read URLs from a CSV file into a list
df = pd.read_csv('urls.csv')
urls = df['url'].tolist() # replace 'url' with your actual column name
start, end = max(0, start), min(len(urls), end)
urls = urls[start:end]
print(f'Fetching {len(urls)} reviews')
data = get_reviews(urls)
print(f'Fetching complete. Inserting into database')
insert_into_db(data, c)
print('Done')
conn.commit()
conn.close()
if __name__ == '__main__':
main(int(argv[1]), int(argv[2]))
|