pitchfork / get_url.py
mattismegevand's picture
init commit
4d5d3e0
raw
history blame
1.78 kB
import requests
import pandas as pd
from concurrent.futures import ThreadPoolExecutor
from bs4 import BeautifulSoup
from os.path import isfile
from sys import argv
def fetch(args):
''' Fetch a single url '''
url, session = args
response = session.get(url)
page_number = url.split('=')[-1]
soup = BeautifulSoup(response.text, 'lxml')
error = soup.find('div', {'class': 'error-page'})
if error:
print(f'Error page: {url} does not exist')
return []
print('.', end='', flush=True)
return [(page_number, f"https://pitchfork.com{e['href']}") for e in soup.find_all('a', {'class': 'review__link'})]
def get_urls(start, end):
''' Return a list of urls from the Pitchfork reviews page '''
urls = [f'https://pitchfork.com/reviews/albums/?page={i}' for i in range(start, end+1)]
reviews = []
session = requests.Session()
with ThreadPoolExecutor(max_workers=5) as executor:
for result in executor.map(fetch, ((url, session) for url in urls)):
reviews.extend(result)
print()
return reviews
def insert_into_df(data):
''' Insert data into a pandas dataframe '''
df = pd.DataFrame(data, columns=['page', 'url'])
df.drop_duplicates(subset='url', keep='first', inplace=True)
return df
def main():
start, end = int(argv[1]), int(argv[2])
print(f'Fetching urls from pages {start} to {end}')
data = get_urls(start, end)
print(f'Fetched {len(data)} urls')
df = insert_into_df(data)
print(f'Writing to urls.csv')
if isfile('urls.csv'):
df_existing = pd.read_csv('urls.csv')
df_combined = pd.concat([df_existing, df])
else:
df_combined = df
df_combined.drop_duplicates(subset='url', keep='first', inplace=True)
df_combined.to_csv('urls.csv', index=False)
print('Done')
if __name__ == '__main__':
main()