news_fr / extract_news /2_extract_news.py
eckendoerffer's picture
Upload 17 files
fc9c13b
raw
history blame
3.36 kB
# -*- coding: utf-8 -*-
"""
News source extractor:
This script is designed to extract the content of news articles from various French media sources.
The URLs of these articles are retrieved from the `base_news` table, where articles marked with
a `step` value of '0' are pending extraction.
To install the necessary packages:
pip install aiohttp mysql-connector-python
Once extracted, the content of each article is saved locally for further processing. This separation
of content fetching and processing is intentional to optimize resource management.
The script operates in batches, processing a defined number of entries (`NB_BY_STEP`) at a time.
After extraction, the `step` value of the processed articles is updated to '1' to indicate completion.
For performance monitoring, the script outputs the processed article IDs, URLs, and calculates the average
processing time per article.
Author : Guillaume Eckendoerffer
Date : 02-10-23
Repository : https://github.com/Eckendoerffer/TorchTrainerFlow/
https://huggingface.co/datasets/eckendoerffer/news_fr
"""
import asyncio
import aiohttp
import time, os
from config import DB_CONFIG
from utils import create_connection, mysqli_return_number, save_to_file
NB_BY_STEP = 20
path = os.getcwd()
async def fetch_and_save(url, id_source):
time_start_item = time.time()
try:
async with aiohttp.ClientSession() as session:
async with session.get(url, timeout=10) as response:
byte_content = await response.read()
try:
text_content = byte_content.decode('utf-8')
except UnicodeDecodeError:
text_content = byte_content.decode('ISO-8859-1')
save_to_file(f"{path}/sources/html_news/{id_source}.txt", text_content)
time_end_item = time.time()
print(f'{id_source}) {time_end_item-time_start_item:.5f} {url}')
except aiohttp.client_exceptions.TooManyRedirects:
print(f"Too many redirects for URL: {url}")
except aiohttp.client_exceptions.ClientConnectorError:
print(f"Failed to connect to URL: {url}")
except Exception as e:
print(f"Unexpected error for URL {url}: {str(e)}")
async def main():
connection = create_connection(DB_CONFIG)
while True:
time_start = time.time()
cursor = connection.cursor()
cursor.execute(f"SELECT `id`, `url` FROM `base_news` WHERE `step`='0' ORDER BY RAND() LIMIT {NB_BY_STEP}")
rows = cursor.fetchall()
cursor.close()
if not rows:
break
tasks = []
for row in rows:
id_source, url = row
cursor = connection.cursor()
cursor.execute(f"UPDATE `base_news` SET `step`='1' WHERE `id`='{id_source}' LIMIT 1")
cursor.close()
tasks.append(fetch_and_save(url.strip(), id_source))
await asyncio.gather(*tasks)
nb_base = mysqli_return_number(connection, "SELECT COUNT(`id`) FROM `base_news` WHERE `step`='0'")
time_elapsed = time.time() - time_start
time_per_item = time_elapsed / NB_BY_STEP
print(f"Remaining: {nb_base} - Time: {time_per_item:.3f}s/url")
#if nb_base < 100: time.sleep(600)
connection.close()
if __name__ == "__main__":
asyncio.run(main())