|
import asyncio |
|
import os |
|
import random |
|
from collections import deque |
|
import ssl |
|
from urllib.parse import urlparse |
|
|
|
import aiohttp |
|
import polars as pl |
|
import aiofiles |
|
from aiohttp import TCPConnector |
|
|
|
|
|
import certifi |
|
|
|
from tenacity import retry, wait_exponential, wait_random_exponential |
|
from tqdm.asyncio import tqdm_asyncio |
|
|
|
|
|
|
|
BATCH_SIZE = 100 |
|
BASE_DOWNLOAD_PATH = "/mnt/jupiter/openalex_extraction/openalex_2" |
|
TIME_OUT = 60 |
|
USER_AGENTS = [ |
|
"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36", |
|
"Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36", |
|
"Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36", |
|
"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Firefox/89.0", |
|
"Mozilla/5.0 (Macintosh; Intel Mac OS X 10.15; rv:89.0) Gecko/20100101 Firefox/89.0", |
|
] |
|
|
|
HEADER = lambda x: { |
|
"User-Agent": random.choice(USER_AGENTS), |
|
"Referer": x, |
|
} |
|
|
|
|
|
async def get_request(session, queries: list, uuid: str) -> tuple[str, bytes | None]: |
|
|
|
ssl_context = ssl.create_default_context() |
|
ssl_context.check_hostname = True |
|
ssl_context.verify_mode = ssl.CERT_REQUIRED |
|
|
|
for query in queries: |
|
try: |
|
async with session.get(url=query, headers=HEADER(query)) as response: |
|
content = await response.read() |
|
uuid = urlparse(uuid).path.replace(r"/", "") |
|
filename = os.path.join(BASE_DOWNLOAD_PATH, f"{uuid}.pdf") |
|
file_number = 1 |
|
while os.path.exists(filename): |
|
filename = os.path.join( |
|
BASE_DOWNLOAD_PATH, f"{uuid}_{file_number}.pdf" |
|
) |
|
file_number += 1 |
|
|
|
|
|
return filename, content |
|
except Exception as e: |
|
print(f"An error occurred with query {query}: {e}") |
|
continue |
|
|
|
|
|
print(f"All queries failed: {queries}") |
|
return "error", None |
|
|
|
|
|
async def get_batched(session, batch, seen: set = None): |
|
if not seen: |
|
seen = set() |
|
tasks = [] |
|
for q in batch: |
|
if q: |
|
task = asyncio.ensure_future(get_request(session, queries=q[0], uuid=q[1])) |
|
tasks.append(task) |
|
return await tqdm_asyncio.gather( |
|
*tasks, desc="Collecting batch", leave=True, position=0 |
|
) |
|
|
|
|
|
async def main(file_loc): |
|
|
|
df = pl.scan_parquet(file_loc) |
|
|
|
total_rows = df.collect().height |
|
|
|
|
|
start_index = 0 |
|
num_rows = total_rows - start_index |
|
|
|
df = df.slice(0, num_rows) |
|
df = ( |
|
df.with_columns(pl.col("pdf_url").str.split(",")) |
|
.select(["identifier", "pdf_url"]) |
|
.collect(streaming=True) |
|
.iter_rows(named=True) |
|
) |
|
|
|
batches = deque() |
|
output = [] |
|
|
|
|
|
ssl_context = ssl.create_default_context() |
|
ssl_context.check_hostname = True |
|
ssl_context.verify_mode = ssl.CERT_REQUIRED |
|
|
|
timeout = aiohttp.ClientTimeout(total=TIME_OUT) |
|
|
|
for row in df: |
|
batches.append((row["pdf_url"], row["identifier"])) |
|
if len(batches) == BATCH_SIZE: |
|
async with aiohttp.ClientSession( |
|
connector=TCPConnector(ssl=ssl_context, limit=50), |
|
timeout=timeout, |
|
) as session: |
|
responses = await get_batched(session, batches) |
|
for filename, content in responses: |
|
if content: |
|
with open(filename, "wb") as f: |
|
f.write(content) |
|
output.append(filename) |
|
batches.clear() |
|
|
|
if batches: |
|
async with aiohttp.ClientSession( |
|
connector=TCPConnector(ssl=ssl_context, limit=50), timeout=timeout |
|
) as session: |
|
responses = await get_batched(session, batches) |
|
print("Saving Batch") |
|
for filename, content in responses: |
|
if content: |
|
with open(filename, "wb") as f: |
|
f.write(content) |
|
output.append(filename) |
|
print("Batch Saved") |
|
|
|
return output |
|
|
|
|
|
if __name__ == "__main__": |
|
FILE_LOCATION = "/mnt/jupiter/openalex_extraction/openalex_extraction_2.parquet" |
|
results = asyncio.run(main(FILE_LOCATION)) |
|
print(results) |