|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
"""Download ColliderML dataset assets and manage README URL updates.""" |
|
|
|
|
|
from __future__ import annotations |
|
|
|
|
|
import asyncio |
|
|
import contextlib |
|
|
import hashlib |
|
|
import logging |
|
|
import os |
|
|
import re |
|
|
import shutil |
|
|
import tempfile |
|
|
from dataclasses import dataclass |
|
|
from pathlib import Path |
|
|
from typing import Annotated, Any, Dict, List, Sequence |
|
|
from urllib.parse import urlparse |
|
|
|
|
|
import aiohttp |
|
|
import pyarrow.parquet as pq |
|
|
import typer |
|
|
import yaml |
|
|
from rich.console import Console |
|
|
from rich.logging import RichHandler |
|
|
from rich.progress import ( |
|
|
BarColumn, |
|
|
DownloadColumn, |
|
|
Progress, |
|
|
TextColumn, |
|
|
TimeElapsedColumn, |
|
|
TransferSpeedColumn, |
|
|
) |
|
|
from rich.table import Table |
|
|
|
|
|
|
|
|
console = Console() |
|
|
app = typer.Typer() |
|
|
|
|
|
|
|
|
logging.basicConfig( |
|
|
level=logging.INFO, |
|
|
format="%(message)s", |
|
|
datefmt="[%X]", |
|
|
handlers=[RichHandler(console=console, show_path=False)], |
|
|
) |
|
|
logger = logging.getLogger(__name__) |
|
|
|
|
|
|
|
|
@app.callback() |
|
|
def main_callback( |
|
|
verbose: Annotated[ |
|
|
bool, |
|
|
typer.Option( |
|
|
"--verbose", |
|
|
"-v", |
|
|
help="Enable verbose logging.", |
|
|
), |
|
|
] = False, |
|
|
) -> None: |
|
|
"""Manage dataset file URLs declared in README front matter.""" |
|
|
if verbose: |
|
|
logging.getLogger().setLevel(logging.DEBUG) |
|
|
logger.debug("Verbose logging enabled") |
|
|
|
|
|
|
|
|
@dataclass |
|
|
class DataFileEntry: |
|
|
config_name: str |
|
|
path: Sequence[str] |
|
|
url: str |
|
|
|
|
|
def parsed(self) -> tuple[str, str]: |
|
|
parsed = urlparse(self.url) |
|
|
return parsed.netloc, parsed.path.lstrip("/") |
|
|
|
|
|
def relative_path(self) -> Path: |
|
|
netloc, remainder = self.parsed() |
|
|
return Path(netloc) / Path(remainder) |
|
|
|
|
|
|
|
|
@dataclass |
|
|
class DownloadResult: |
|
|
entry: DataFileEntry |
|
|
path: Path |
|
|
success: bool |
|
|
skipped: bool |
|
|
error: Exception | None = None |
|
|
order: int = 0 |
|
|
|
|
|
|
|
|
@dataclass |
|
|
class VerifyResult: |
|
|
entry: DataFileEntry |
|
|
ok: bool |
|
|
status: int | None |
|
|
detail: str = "" |
|
|
order: int = 0 |
|
|
|
|
|
|
|
|
@dataclass |
|
|
class ChecksumResult: |
|
|
entry: DataFileEntry |
|
|
filename: str |
|
|
computed_hash: str | None |
|
|
expected_hash: str | None |
|
|
matches: bool |
|
|
success: bool |
|
|
error: Exception | None = None |
|
|
order: int = 0 |
|
|
|
|
|
|
|
|
@dataclass |
|
|
class SchemaInfo: |
|
|
config_name: str |
|
|
filename: str |
|
|
columns: List[tuple[str, str]] |
|
|
num_rows: int | None |
|
|
success: bool |
|
|
error: Exception | None = None |
|
|
|
|
|
|
|
|
def read_front_matter(readme_path: Path) -> tuple[str, str]: |
|
|
text = readme_path.read_text(encoding="utf-8") |
|
|
if not text.startswith("---\n"): |
|
|
raise ValueError("README.md does not start with YAML front matter.") |
|
|
try: |
|
|
front_matter, body = text[4:].split("\n---\n", 1) |
|
|
except ValueError as exc: |
|
|
raise ValueError("README.md front matter is not terminated by '---'.") from exc |
|
|
return front_matter, body |
|
|
|
|
|
|
|
|
def load_data_file_entries(front_matter_text: str) -> List[DataFileEntry]: |
|
|
data = yaml.safe_load(front_matter_text) |
|
|
configs = data.get("configs", []) if isinstance(data, dict) else [] |
|
|
entries: List[DataFileEntry] = [] |
|
|
|
|
|
def _walk(value: Any, path: list[str]) -> None: |
|
|
if value is None: |
|
|
return |
|
|
if isinstance(value, str): |
|
|
entries.append( |
|
|
DataFileEntry(config_name=current_config, path=tuple(path), url=value) |
|
|
) |
|
|
elif isinstance(value, list): |
|
|
for idx, item in enumerate(value): |
|
|
_walk(item, [*path, str(idx)]) |
|
|
elif isinstance(value, dict): |
|
|
for key, item in value.items(): |
|
|
_walk(item, [*path, str(key)]) |
|
|
|
|
|
for config in configs: |
|
|
if not isinstance(config, dict): |
|
|
continue |
|
|
current_config = str(config.get("config_name", "<unnamed>")) |
|
|
_walk(config.get("data_files"), ["data_files"]) |
|
|
|
|
|
return entries |
|
|
|
|
|
|
|
|
def replace_once(text: str, old: str, new: str) -> str: |
|
|
index = text.find(old) |
|
|
if index == -1: |
|
|
raise ValueError(f"Value '{old}' was not found in front matter.") |
|
|
return f"{text[:index]}{new}{text[index + len(old):]}" |
|
|
|
|
|
|
|
|
def build_rewritten_url( |
|
|
result: DownloadResult, output_dir: Path, base_url: str |
|
|
) -> str: |
|
|
"""Build new URL by replacing output_dir with base_url in the file path. |
|
|
|
|
|
For example: |
|
|
- File path: /data/output/particles/file.parquet |
|
|
- Output dir: /data/output |
|
|
- Base URL: https://example.com/files |
|
|
- Result: https://example.com/files/particles/file.parquet |
|
|
""" |
|
|
relative_path = result.path.relative_to(output_dir) |
|
|
|
|
|
base_url = base_url.rstrip("/") |
|
|
return f"{base_url}/{relative_path.as_posix()}" |
|
|
|
|
|
|
|
|
def resolve_destination(entry: DataFileEntry, output_dir: Path) -> Path: |
|
|
_, remote_path = entry.parsed() |
|
|
filename = Path(remote_path).name or remote_path |
|
|
base_dir = output_dir / entry.config_name |
|
|
return (base_dir / filename).resolve() |
|
|
|
|
|
|
|
|
def move_with_progress(src: Path, dst: Path, progress: Progress, task_id: int) -> None: |
|
|
"""Move file with progress tracking for cross-filesystem moves.""" |
|
|
|
|
|
try: |
|
|
os.rename(src, dst) |
|
|
return |
|
|
except OSError: |
|
|
|
|
|
pass |
|
|
|
|
|
|
|
|
file_size = src.stat().st_size |
|
|
progress.update(task_id, total=file_size) |
|
|
|
|
|
|
|
|
buffer_size = 8 << 20 |
|
|
with src.open("rb") as fsrc, dst.open("wb") as fdst: |
|
|
copied = 0 |
|
|
while True: |
|
|
buf = fsrc.read(buffer_size) |
|
|
if not buf: |
|
|
break |
|
|
fdst.write(buf) |
|
|
copied += len(buf) |
|
|
progress.update(task_id, completed=copied) |
|
|
|
|
|
|
|
|
fdst.flush() |
|
|
os.fsync(fdst.fileno()) |
|
|
|
|
|
|
|
|
if dst.stat().st_size != file_size: |
|
|
raise IOError( |
|
|
f"File size mismatch after copy: expected {file_size}, got {dst.stat().st_size}" |
|
|
) |
|
|
|
|
|
|
|
|
src.unlink() |
|
|
|
|
|
|
|
|
async def download_one( |
|
|
entry: DataFileEntry, |
|
|
output_dir: Path, |
|
|
session: aiohttp.ClientSession, |
|
|
semaphore: asyncio.Semaphore, |
|
|
skip_existing: bool, |
|
|
progress: Progress, |
|
|
order: int, |
|
|
staging_dir: Path | None = None, |
|
|
) -> DownloadResult: |
|
|
_, remote_path = entry.parsed() |
|
|
filename = Path(remote_path).name or remote_path |
|
|
terse_name = (filename[:32] + "…") if len(filename) > 33 else filename |
|
|
description = f"{entry.config_name}: {terse_name}" |
|
|
async with semaphore: |
|
|
task_id: int | None = None |
|
|
destination = resolve_destination(entry, output_dir) |
|
|
|
|
|
|
|
|
if staging_dir: |
|
|
download_dest = resolve_destination(entry, staging_dir) |
|
|
tmp_path = download_dest.parent / f"{download_dest.name}.part" |
|
|
else: |
|
|
download_dest = destination |
|
|
tmp_path = destination.parent / f"{destination.name}.part" |
|
|
|
|
|
try: |
|
|
destination.parent.mkdir(parents=True, exist_ok=True) |
|
|
download_dest.parent.mkdir(parents=True, exist_ok=True) |
|
|
task_id = progress.add_task(description, total=0, start=False) |
|
|
progress.start_task(task_id) |
|
|
async with session.get(entry.url) as response: |
|
|
response.raise_for_status() |
|
|
total_bytes = response.content_length or 0 |
|
|
|
|
|
|
|
|
if skip_existing and destination.exists(): |
|
|
local_size = destination.stat().st_size |
|
|
if total_bytes and local_size == total_bytes: |
|
|
if task_id is not None: |
|
|
progress.remove_task(task_id) |
|
|
task_id = None |
|
|
logger.info( |
|
|
f"Skipped {destination.name} ({total_bytes:,} bytes) - already exists with correct size" |
|
|
) |
|
|
return DownloadResult( |
|
|
entry=entry, |
|
|
path=destination, |
|
|
success=True, |
|
|
skipped=True, |
|
|
order=order, |
|
|
) |
|
|
else: |
|
|
|
|
|
if not total_bytes: |
|
|
reason = f"remote size unknown, local size is {local_size:,} bytes" |
|
|
else: |
|
|
reason = f"size mismatch (local: {local_size:,} bytes, remote: {total_bytes:,} bytes)" |
|
|
logger.info(f"Downloading {destination.name} - {reason}") |
|
|
else: |
|
|
|
|
|
if not destination.exists(): |
|
|
size_info = f" ({total_bytes:,} bytes)" if total_bytes else "" |
|
|
logger.info( |
|
|
f"Downloading {destination.name}{size_info} - file not found locally" |
|
|
) |
|
|
else: |
|
|
size_info = f" ({total_bytes:,} bytes)" if total_bytes else "" |
|
|
logger.info( |
|
|
f"Downloading {destination.name}{size_info} - skip_existing is disabled" |
|
|
) |
|
|
|
|
|
if total_bytes: |
|
|
progress.update(task_id, total=total_bytes) |
|
|
with tmp_path.open("wb") as handle: |
|
|
async for chunk in response.content.iter_chunked(1 << 17): |
|
|
handle.write(chunk) |
|
|
progress.update(task_id, advance=len(chunk)) |
|
|
|
|
|
|
|
|
handle.flush() |
|
|
os.fsync(handle.fileno()) |
|
|
|
|
|
|
|
|
tmp_path.rename(download_dest) |
|
|
|
|
|
|
|
|
if staging_dir: |
|
|
logger.info( |
|
|
f"Moving {download_dest.name} from staging to {destination.parent.name}/" |
|
|
) |
|
|
|
|
|
if task_id is not None: |
|
|
progress.update( |
|
|
task_id, description=f"{entry.config_name}: Moving {terse_name}" |
|
|
) |
|
|
await asyncio.to_thread( |
|
|
move_with_progress, download_dest, destination, progress, task_id |
|
|
) |
|
|
|
|
|
return DownloadResult( |
|
|
entry=entry, |
|
|
path=destination, |
|
|
success=True, |
|
|
skipped=False, |
|
|
order=order, |
|
|
) |
|
|
except Exception as exc: |
|
|
|
|
|
logger.error(f"Failed to download {entry.config_name}/{filename}: {exc}") |
|
|
with contextlib.suppress(FileNotFoundError): |
|
|
tmp_path.unlink() |
|
|
return DownloadResult( |
|
|
entry=entry, |
|
|
path=destination, |
|
|
success=False, |
|
|
skipped=False, |
|
|
error=exc, |
|
|
order=order, |
|
|
) |
|
|
finally: |
|
|
if task_id is not None: |
|
|
progress.remove_task(task_id) |
|
|
|
|
|
|
|
|
async def perform_downloads( |
|
|
entries: Sequence[DataFileEntry], |
|
|
output_dir: Path, |
|
|
max_concurrency: int, |
|
|
timeout: float, |
|
|
skip_existing: bool, |
|
|
use_staging: bool = False, |
|
|
) -> List[DownloadResult]: |
|
|
if not entries: |
|
|
return [] |
|
|
|
|
|
with contextlib.ExitStack() as stack: |
|
|
|
|
|
staging_dir: Path | None = None |
|
|
if use_staging: |
|
|
staging_tmp = stack.enter_context( |
|
|
tempfile.TemporaryDirectory(prefix="dataset_staging_") |
|
|
) |
|
|
staging_dir = Path(staging_tmp) |
|
|
logger.info(f"Using staging directory: {staging_dir}") |
|
|
|
|
|
semaphore = asyncio.Semaphore(max_concurrency) |
|
|
results: List[DownloadResult] = [] |
|
|
timeout_cfg = aiohttp.ClientTimeout(total=timeout) |
|
|
progress = Progress( |
|
|
TextColumn("{task.description}"), |
|
|
BarColumn(bar_width=None), |
|
|
DownloadColumn(), |
|
|
TransferSpeedColumn(), |
|
|
TimeElapsedColumn(), |
|
|
console=console, |
|
|
) |
|
|
|
|
|
async with aiohttp.ClientSession(timeout=timeout_cfg) as session: |
|
|
with progress: |
|
|
tasks: list[asyncio.Task[DownloadResult]] = [] |
|
|
for order, entry in enumerate(entries): |
|
|
task = asyncio.create_task( |
|
|
download_one( |
|
|
entry=entry, |
|
|
output_dir=output_dir, |
|
|
session=session, |
|
|
semaphore=semaphore, |
|
|
skip_existing=skip_existing, |
|
|
progress=progress, |
|
|
order=order, |
|
|
staging_dir=staging_dir, |
|
|
) |
|
|
) |
|
|
tasks.append(task) |
|
|
|
|
|
for future in asyncio.as_completed(tasks): |
|
|
result = await future |
|
|
results.append(result) |
|
|
|
|
|
results.sort(key=lambda item: item.order) |
|
|
return results |
|
|
|
|
|
|
|
|
async def get_remote_file_size( |
|
|
entry: DataFileEntry, |
|
|
session: aiohttp.ClientSession, |
|
|
semaphore: asyncio.Semaphore, |
|
|
) -> int | None: |
|
|
"""Get remote file size via HEAD request.""" |
|
|
async with semaphore: |
|
|
try: |
|
|
async with session.head(entry.url, allow_redirects=True) as response: |
|
|
if response.status < 400: |
|
|
return response.content_length |
|
|
except Exception: |
|
|
pass |
|
|
return None |
|
|
|
|
|
|
|
|
async def fetch_remote_sizes( |
|
|
entries: Sequence[DataFileEntry], |
|
|
max_concurrency: int, |
|
|
timeout: float, |
|
|
) -> Dict[str, int | None]: |
|
|
"""Fetch remote file sizes for all entries.""" |
|
|
semaphore = asyncio.Semaphore(max_concurrency) |
|
|
timeout_cfg = aiohttp.ClientTimeout(total=timeout) |
|
|
sizes: Dict[str, int | None] = {} |
|
|
|
|
|
progress = Progress( |
|
|
TextColumn("Fetching remote file sizes..."), |
|
|
BarColumn(), |
|
|
TextColumn("{task.completed}/{task.total}"), |
|
|
TimeElapsedColumn(), |
|
|
console=console, |
|
|
) |
|
|
|
|
|
async with aiohttp.ClientSession(timeout=timeout_cfg) as session: |
|
|
with progress: |
|
|
task_id = progress.add_task("Fetching sizes", total=len(entries)) |
|
|
tasks = { |
|
|
entry.url: asyncio.create_task( |
|
|
get_remote_file_size(entry, session, semaphore) |
|
|
) |
|
|
for entry in entries |
|
|
} |
|
|
for url, task in tasks.items(): |
|
|
size = await task |
|
|
sizes[url] = size |
|
|
progress.advance(task_id) |
|
|
|
|
|
return sizes |
|
|
|
|
|
|
|
|
async def verify_one( |
|
|
entry: DataFileEntry, |
|
|
session: aiohttp.ClientSession, |
|
|
semaphore: asyncio.Semaphore, |
|
|
order: int, |
|
|
) -> VerifyResult: |
|
|
async with semaphore: |
|
|
last_error: str = "" |
|
|
for method in ("HEAD", "GET"): |
|
|
try: |
|
|
async with session.request( |
|
|
method, entry.url, allow_redirects=True |
|
|
) as response: |
|
|
status = response.status |
|
|
if status < 400: |
|
|
return VerifyResult( |
|
|
entry=entry, ok=True, status=status, order=order |
|
|
) |
|
|
if method == "HEAD" and status in {405, 501}: |
|
|
last_error = f"{method} returned {status}; retrying with GET." |
|
|
continue |
|
|
return VerifyResult( |
|
|
entry=entry, |
|
|
ok=False, |
|
|
status=status, |
|
|
detail=f"{method} -> {status}", |
|
|
order=order, |
|
|
) |
|
|
except Exception as exc: |
|
|
last_error = str(exc) |
|
|
return VerifyResult( |
|
|
entry=entry, |
|
|
ok=False, |
|
|
status=None, |
|
|
detail=last_error or "Unknown error", |
|
|
order=order, |
|
|
) |
|
|
|
|
|
|
|
|
async def perform_verification( |
|
|
entries: Sequence[DataFileEntry], |
|
|
max_concurrency: int, |
|
|
timeout: float, |
|
|
) -> List[VerifyResult]: |
|
|
if not entries: |
|
|
return [] |
|
|
|
|
|
semaphore = asyncio.Semaphore(max_concurrency) |
|
|
timeout_cfg = aiohttp.ClientTimeout(total=timeout) |
|
|
results: List[VerifyResult] = [] |
|
|
progress = Progress( |
|
|
TextColumn("{task.description}"), |
|
|
BarColumn(), |
|
|
TextColumn("{task.completed}/{task.total}"), |
|
|
TimeElapsedColumn(), |
|
|
console=console, |
|
|
) |
|
|
|
|
|
async with aiohttp.ClientSession(timeout=timeout_cfg) as session: |
|
|
with progress: |
|
|
task_id = progress.add_task("Verifying dataset URLs", total=len(entries)) |
|
|
tasks = [ |
|
|
asyncio.create_task(verify_one(entry, session, semaphore, order=order)) |
|
|
for order, entry in enumerate(entries) |
|
|
] |
|
|
for future in asyncio.as_completed(tasks): |
|
|
result = await future |
|
|
results.append(result) |
|
|
progress.advance(task_id) |
|
|
|
|
|
results.sort(key=lambda item: item.order) |
|
|
return results |
|
|
|
|
|
|
|
|
async def compute_file_hash( |
|
|
entry: DataFileEntry, |
|
|
session: aiohttp.ClientSession, |
|
|
semaphore: asyncio.Semaphore, |
|
|
progress: Progress, |
|
|
order: int, |
|
|
) -> ChecksumResult: |
|
|
"""Download file to temp location and compute SHA256 hash.""" |
|
|
|
|
|
_, remote_path = entry.parsed() |
|
|
filename = Path(remote_path).name or remote_path |
|
|
terse_name = (filename[:32] + "…") if len(filename) > 33 else filename |
|
|
description = f"{entry.config_name}: {terse_name}" |
|
|
|
|
|
async with semaphore: |
|
|
task_id: int | None = None |
|
|
try: |
|
|
task_id = progress.add_task(description, total=0, start=False) |
|
|
progress.start_task(task_id) |
|
|
|
|
|
async with session.get(entry.url) as response: |
|
|
response.raise_for_status() |
|
|
total_bytes = response.content_length or 0 |
|
|
|
|
|
if total_bytes: |
|
|
progress.update(task_id, total=total_bytes) |
|
|
|
|
|
hasher = hashlib.sha256() |
|
|
async for chunk in response.content.iter_chunked(1 << 17): |
|
|
hasher.update(chunk) |
|
|
progress.update(task_id, advance=len(chunk)) |
|
|
|
|
|
computed_hash = hasher.hexdigest() |
|
|
return ChecksumResult( |
|
|
entry=entry, |
|
|
filename=filename, |
|
|
computed_hash=computed_hash, |
|
|
expected_hash=None, |
|
|
matches=False, |
|
|
success=True, |
|
|
order=order, |
|
|
) |
|
|
except Exception as exc: |
|
|
return ChecksumResult( |
|
|
entry=entry, |
|
|
filename=filename, |
|
|
computed_hash=None, |
|
|
expected_hash=None, |
|
|
matches=False, |
|
|
success=False, |
|
|
error=exc, |
|
|
order=order, |
|
|
) |
|
|
finally: |
|
|
if task_id is not None: |
|
|
progress.remove_task(task_id) |
|
|
|
|
|
|
|
|
async def perform_checksum_verification( |
|
|
entries: Sequence[DataFileEntry], |
|
|
expected_hashes_by_config: Dict[str, Dict[str, str]], |
|
|
max_concurrency: int, |
|
|
timeout: float, |
|
|
) -> List[ChecksumResult]: |
|
|
"""Download files and verify their checksums. |
|
|
|
|
|
expected_hashes_by_config: dict mapping config_name -> (filename -> hash) |
|
|
""" |
|
|
if not entries: |
|
|
return [] |
|
|
|
|
|
semaphore = asyncio.Semaphore(max_concurrency) |
|
|
timeout_cfg = aiohttp.ClientTimeout(total=timeout) |
|
|
results: List[ChecksumResult] = [] |
|
|
progress = Progress( |
|
|
TextColumn("{task.description}"), |
|
|
BarColumn(), |
|
|
DownloadColumn(), |
|
|
TransferSpeedColumn(), |
|
|
TimeElapsedColumn(), |
|
|
console=console, |
|
|
) |
|
|
|
|
|
async with aiohttp.ClientSession(timeout=timeout_cfg) as session: |
|
|
with progress: |
|
|
tasks = [ |
|
|
asyncio.create_task( |
|
|
compute_file_hash( |
|
|
entry, session, semaphore, order=order, progress=progress |
|
|
) |
|
|
) |
|
|
for order, entry in enumerate(entries) |
|
|
] |
|
|
for future in asyncio.as_completed(tasks): |
|
|
result = await future |
|
|
|
|
|
config_hashes = expected_hashes_by_config.get( |
|
|
result.entry.config_name, {} |
|
|
) |
|
|
result.expected_hash = config_hashes.get(result.filename) |
|
|
if result.success and result.expected_hash: |
|
|
result.matches = result.computed_hash == result.expected_hash |
|
|
results.append(result) |
|
|
|
|
|
results.sort(key=lambda item: item.order) |
|
|
return results |
|
|
|
|
|
|
|
|
def load_checksums_for_config(checksum_dir: Path, config_name: str) -> Dict[str, str]: |
|
|
"""Load expected hashes from SHA256SUM-style file for a specific config. |
|
|
|
|
|
Returns a dict mapping filename -> hash. |
|
|
""" |
|
|
checksum_file = checksum_dir / f"{config_name}.sha256" |
|
|
if not checksum_file.exists(): |
|
|
return {} |
|
|
|
|
|
hashes: Dict[str, str] = {} |
|
|
try: |
|
|
with checksum_file.open("r", encoding="utf-8") as f: |
|
|
for line in f: |
|
|
line = line.strip() |
|
|
if not line or line.startswith("#"): |
|
|
continue |
|
|
|
|
|
parts = line.split(None, 1) |
|
|
if len(parts) == 2: |
|
|
hash_value, filename = parts |
|
|
hashes[filename] = hash_value |
|
|
except OSError as exc: |
|
|
logger.warning(f"Failed to load checksum file {checksum_file}: {exc}") |
|
|
|
|
|
return hashes |
|
|
|
|
|
|
|
|
def save_checksums_for_config( |
|
|
checksum_dir: Path, config_name: str, file_hashes: Dict[str, str] |
|
|
) -> None: |
|
|
"""Save hashes to SHA256SUM-style file for a specific config. |
|
|
|
|
|
file_hashes: dict mapping filename -> hash |
|
|
""" |
|
|
checksum_dir.mkdir(parents=True, exist_ok=True) |
|
|
checksum_file = checksum_dir / f"{config_name}.sha256" |
|
|
|
|
|
with checksum_file.open("w", encoding="utf-8") as f: |
|
|
for filename in sorted(file_hashes.keys()): |
|
|
hash_value = file_hashes[filename] |
|
|
f.write(f"{hash_value} {filename}\n") |
|
|
|
|
|
logger.info(f"Saved {len(file_hashes)} checksums to {checksum_file}") |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
@app.command() |
|
|
def download( |
|
|
output_dir: Annotated[ |
|
|
Path, |
|
|
typer.Option( |
|
|
"--output-dir", |
|
|
"-o", |
|
|
help="Directory where files will be stored.", |
|
|
resolve_path=True, |
|
|
), |
|
|
], |
|
|
readme_path: Annotated[ |
|
|
Path, |
|
|
typer.Option( |
|
|
"--readme-path", |
|
|
"-r", |
|
|
help="Path to the README file with YAML front matter.", |
|
|
exists=True, |
|
|
resolve_path=True, |
|
|
dir_okay=False, |
|
|
), |
|
|
] = Path("README.md"), |
|
|
max_concurrency: Annotated[ |
|
|
int, |
|
|
typer.Option( |
|
|
"--max-concurrency", |
|
|
"-c", |
|
|
min=1, |
|
|
show_default=True, |
|
|
help="Maximum number of concurrent downloads.", |
|
|
), |
|
|
] = 4, |
|
|
timeout: Annotated[ |
|
|
float, |
|
|
typer.Option( |
|
|
"--timeout", |
|
|
min=1.0, |
|
|
show_default=True, |
|
|
help="Request timeout in seconds.", |
|
|
), |
|
|
] = 600.0, |
|
|
dry_run: Annotated[ |
|
|
bool, |
|
|
typer.Option( |
|
|
"--dry-run/--no-dry-run", |
|
|
show_default=True, |
|
|
help="Preview downloads and README rewrites without performing any changes.", |
|
|
), |
|
|
] = False, |
|
|
skip_existing: Annotated[ |
|
|
bool, |
|
|
typer.Option( |
|
|
"--skip-existing/--no-skip-existing", |
|
|
show_default=True, |
|
|
help="Skip downloading files that already exist locally.", |
|
|
), |
|
|
] = True, |
|
|
use_staging: Annotated[ |
|
|
bool, |
|
|
typer.Option( |
|
|
"--stage/--no-stage", |
|
|
show_default=True, |
|
|
help="Download to a temporary staging directory first, then move to final destination.", |
|
|
), |
|
|
] = False, |
|
|
clean_invalid: Annotated[ |
|
|
bool, |
|
|
typer.Option( |
|
|
"--clean-invalid", |
|
|
help="Delete files with incorrect sizes but don't download. Useful for cleaning up failed downloads.", |
|
|
), |
|
|
] = False, |
|
|
) -> None: |
|
|
front_matter_text, body_text = read_front_matter(readme_path) |
|
|
entries = load_data_file_entries(front_matter_text) |
|
|
if not entries: |
|
|
logger.warning("No data_files entries found in README front matter.") |
|
|
raise typer.Exit(code=0) |
|
|
|
|
|
logger.info(f"Found {len(entries)} data file URLs across {readme_path}.") |
|
|
output_dir = output_dir.resolve() |
|
|
|
|
|
if clean_invalid: |
|
|
|
|
|
logger.info("Fetching remote file sizes to check for invalid files...") |
|
|
remote_sizes = asyncio.run( |
|
|
fetch_remote_sizes( |
|
|
entries=entries, |
|
|
max_concurrency=max_concurrency, |
|
|
timeout=timeout, |
|
|
) |
|
|
) |
|
|
|
|
|
|
|
|
deleted = 0 |
|
|
skipped = 0 |
|
|
missing = 0 |
|
|
size_unknown = 0 |
|
|
|
|
|
table = Table(title="Cleaning invalid files") |
|
|
table.add_column("Config") |
|
|
table.add_column("File") |
|
|
table.add_column("Status") |
|
|
table.add_column("Local size") |
|
|
table.add_column("Remote size") |
|
|
|
|
|
for entry in entries: |
|
|
destination = resolve_destination(entry, output_dir) |
|
|
_, remote_path = entry.parsed() |
|
|
filename = Path(remote_path).name or remote_path |
|
|
|
|
|
if not destination.exists(): |
|
|
missing += 1 |
|
|
table.add_row( |
|
|
entry.config_name, |
|
|
filename, |
|
|
"[dim]Not found[/dim]", |
|
|
"-", |
|
|
"-", |
|
|
) |
|
|
continue |
|
|
|
|
|
local_size = destination.stat().st_size |
|
|
remote_size = remote_sizes.get(entry.url) |
|
|
|
|
|
if not remote_size: |
|
|
size_unknown += 1 |
|
|
table.add_row( |
|
|
entry.config_name, |
|
|
filename, |
|
|
"[cyan]Skipped (remote size unknown)[/cyan]", |
|
|
f"{local_size:,}", |
|
|
"Unknown", |
|
|
) |
|
|
continue |
|
|
|
|
|
if local_size == remote_size: |
|
|
skipped += 1 |
|
|
table.add_row( |
|
|
entry.config_name, |
|
|
filename, |
|
|
"[green]Valid[/green]", |
|
|
f"{local_size:,}", |
|
|
f"{remote_size:,}", |
|
|
) |
|
|
else: |
|
|
|
|
|
destination.unlink() |
|
|
deleted += 1 |
|
|
logger.warning( |
|
|
f"Deleted {destination} (size mismatch: {local_size:,} != {remote_size:,})" |
|
|
) |
|
|
table.add_row( |
|
|
entry.config_name, |
|
|
filename, |
|
|
"[red]Deleted (size mismatch)[/red]", |
|
|
f"{local_size:,}", |
|
|
f"{remote_size:,}", |
|
|
) |
|
|
|
|
|
console.print(table) |
|
|
logger.info( |
|
|
f"Summary: {deleted} deleted, {skipped} valid, {missing} missing, {size_unknown} unknown size" |
|
|
) |
|
|
return |
|
|
|
|
|
if dry_run: |
|
|
|
|
|
logger.info("Fetching remote file sizes for dry-run preview...") |
|
|
remote_sizes = asyncio.run( |
|
|
fetch_remote_sizes( |
|
|
entries=entries, |
|
|
max_concurrency=max_concurrency, |
|
|
timeout=timeout, |
|
|
) |
|
|
) |
|
|
|
|
|
preview = Table(title="Download plan (dry-run)") |
|
|
preview.add_column("Config") |
|
|
preview.add_column("Local file", overflow="fold") |
|
|
preview.add_column("Status") |
|
|
preview.add_column("Local size") |
|
|
preview.add_column("Remote size") |
|
|
preview.add_column("Source URL", overflow="fold") |
|
|
|
|
|
for order, entry in enumerate(entries): |
|
|
destination = resolve_destination(entry, output_dir) |
|
|
try: |
|
|
relative = destination.relative_to(output_dir) |
|
|
except ValueError: |
|
|
relative = Path(destination) |
|
|
|
|
|
|
|
|
local_exists = destination.exists() |
|
|
local_size = destination.stat().st_size if local_exists else None |
|
|
remote_size = remote_sizes.get(entry.url) |
|
|
|
|
|
|
|
|
if not local_exists: |
|
|
status = "[yellow]Will download[/yellow]" |
|
|
local_size_str = "-" |
|
|
elif not skip_existing: |
|
|
status = "[red]Will overwrite[/red]" |
|
|
local_size_str = f"{local_size:,}" |
|
|
elif remote_size and local_size == remote_size: |
|
|
status = "[green]Will skip (exists)[/green]" |
|
|
local_size_str = f"{local_size:,}" |
|
|
elif not remote_size: |
|
|
status = "[cyan]Will skip (exists, size unknown)[/cyan]" |
|
|
local_size_str = f"{local_size:,}" |
|
|
else: |
|
|
|
|
|
status = "[red]Will overwrite (size mismatch)[/red]" |
|
|
local_size_str = f"{local_size:,}" |
|
|
|
|
|
remote_size_str = f"{remote_size:,}" if remote_size else "Unknown" |
|
|
|
|
|
row_data = [ |
|
|
entry.config_name, |
|
|
relative.as_posix(), |
|
|
status, |
|
|
local_size_str, |
|
|
remote_size_str, |
|
|
entry.url, |
|
|
] |
|
|
|
|
|
preview.add_row(*row_data) |
|
|
|
|
|
console.print(preview) |
|
|
|
|
|
|
|
|
will_download = 0 |
|
|
will_skip = 0 |
|
|
will_overwrite = 0 |
|
|
|
|
|
for entry in entries: |
|
|
dest = resolve_destination(entry, output_dir) |
|
|
local_exists = dest.exists() |
|
|
local_size = dest.stat().st_size if local_exists else None |
|
|
remote_size = remote_sizes.get(entry.url) |
|
|
|
|
|
if not local_exists: |
|
|
will_download += 1 |
|
|
elif not skip_existing: |
|
|
will_overwrite += 1 |
|
|
elif remote_size and local_size == remote_size: |
|
|
will_skip += 1 |
|
|
elif not remote_size: |
|
|
will_skip += 1 |
|
|
else: |
|
|
will_overwrite += 1 |
|
|
|
|
|
logger.info( |
|
|
f"Summary: {will_download} new, {will_skip} will skip, {will_overwrite} will overwrite" |
|
|
) |
|
|
logger.info("Dry run: no files downloaded and README left unchanged.") |
|
|
return |
|
|
|
|
|
output_dir.mkdir(parents=True, exist_ok=True) |
|
|
|
|
|
results = asyncio.run( |
|
|
perform_downloads( |
|
|
entries=entries, |
|
|
output_dir=output_dir, |
|
|
max_concurrency=max_concurrency, |
|
|
timeout=timeout, |
|
|
skip_existing=skip_existing, |
|
|
use_staging=use_staging, |
|
|
) |
|
|
) |
|
|
|
|
|
successes = sum(1 for item in results if item.success) |
|
|
failures = len(results) - successes |
|
|
skipped = sum(1 for item in results if item.skipped) |
|
|
logger.info(f"{successes} succeeded, {failures} failed, {skipped} skipped.") |
|
|
|
|
|
if failures: |
|
|
for item in results: |
|
|
if not item.success: |
|
|
logger.error(f"Error: {item.entry.url} -> {item.error}") |
|
|
raise typer.Exit(code=1) |
|
|
|
|
|
|
|
|
@app.command() |
|
|
def rewrite( |
|
|
output_dir: Annotated[ |
|
|
Path, |
|
|
typer.Option( |
|
|
"--output-dir", |
|
|
"-o", |
|
|
help="Directory where downloaded files are stored.", |
|
|
resolve_path=True, |
|
|
), |
|
|
], |
|
|
base_url: Annotated[ |
|
|
str, |
|
|
typer.Option( |
|
|
"--base-url", |
|
|
"-u", |
|
|
help=( |
|
|
"Base URL for rewriting URLs in README. Files under output_dir will be rewritten " |
|
|
"to use this base URL. Example: if output_dir=/data/files and base_url=https://example.com/dataset, " |
|
|
"then /data/files/particles/file.parquet becomes https://example.com/dataset/particles/file.parquet" |
|
|
), |
|
|
), |
|
|
], |
|
|
readme_path: Annotated[ |
|
|
Path, |
|
|
typer.Option( |
|
|
"--readme-path", |
|
|
"-r", |
|
|
help="Path to the README file with YAML front matter.", |
|
|
exists=True, |
|
|
resolve_path=True, |
|
|
dir_okay=False, |
|
|
), |
|
|
] = Path("README.md"), |
|
|
dry_run: Annotated[ |
|
|
bool, |
|
|
typer.Option( |
|
|
"--dry-run/--no-dry-run", |
|
|
show_default=True, |
|
|
help="Preview URL rewrites without modifying the README.", |
|
|
), |
|
|
] = False, |
|
|
skip_missing: Annotated[ |
|
|
bool, |
|
|
typer.Option( |
|
|
"--skip-missing/--no-skip-missing", |
|
|
show_default=True, |
|
|
help="Skip rewriting URLs for files that don't exist locally.", |
|
|
), |
|
|
] = True, |
|
|
) -> None: |
|
|
"""Rewrite URLs in README to point to local files served at a base URL.""" |
|
|
front_matter_text, body_text = read_front_matter(readme_path) |
|
|
entries = load_data_file_entries(front_matter_text) |
|
|
if not entries: |
|
|
logger.warning("No data_files entries found in README front matter.") |
|
|
raise typer.Exit(code=0) |
|
|
|
|
|
logger.info(f"Found {len(entries)} data file URLs in {readme_path}.") |
|
|
output_dir = output_dir.resolve() |
|
|
|
|
|
|
|
|
results: List[DownloadResult] = [] |
|
|
missing_files = [] |
|
|
|
|
|
for order, entry in enumerate(entries): |
|
|
destination = resolve_destination(entry, output_dir) |
|
|
if destination.exists(): |
|
|
results.append( |
|
|
DownloadResult( |
|
|
entry=entry, |
|
|
path=destination, |
|
|
success=True, |
|
|
skipped=False, |
|
|
order=order, |
|
|
) |
|
|
) |
|
|
else: |
|
|
missing_files.append((entry, destination)) |
|
|
|
|
|
if missing_files: |
|
|
if not skip_missing: |
|
|
logger.error(f"Found {len(missing_files)} missing files:") |
|
|
for entry, dest in missing_files: |
|
|
logger.error(f" {entry.config_name}: {dest}") |
|
|
logger.error( |
|
|
"Use --skip-missing to rewrite URLs only for existing files, " |
|
|
"or download the missing files first." |
|
|
) |
|
|
raise typer.Exit(code=1) |
|
|
else: |
|
|
logger.warning( |
|
|
f"Skipping {len(missing_files)} missing files " |
|
|
f"(rewriting {len(results)} existing files)" |
|
|
) |
|
|
|
|
|
if not results: |
|
|
logger.error("No files found to rewrite URLs for.") |
|
|
raise typer.Exit(code=1) |
|
|
|
|
|
|
|
|
table = Table(title="URL rewrite preview" if dry_run else "README URL updates") |
|
|
table.add_column("Config") |
|
|
table.add_column("Local file", overflow="fold") |
|
|
table.add_column("Old URL", overflow="fold") |
|
|
table.add_column("New URL", overflow="fold") |
|
|
|
|
|
replacements: list[tuple[DownloadResult, str]] = [] |
|
|
for result in results: |
|
|
relative_path = result.path.relative_to(output_dir) |
|
|
new_url = build_rewritten_url(result, output_dir, base_url) |
|
|
replacements.append((result, new_url)) |
|
|
table.add_row( |
|
|
result.entry.config_name, |
|
|
relative_path.as_posix(), |
|
|
result.entry.url, |
|
|
new_url, |
|
|
) |
|
|
|
|
|
console.print(table) |
|
|
|
|
|
if dry_run: |
|
|
logger.info( |
|
|
f"Dry run: would rewrite {len(results)} URLs. " |
|
|
"Remove --dry-run to apply changes." |
|
|
) |
|
|
return |
|
|
|
|
|
|
|
|
updated_front = front_matter_text |
|
|
for result, new_url in replacements: |
|
|
updated_front = replace_once(updated_front, result.entry.url, new_url) |
|
|
|
|
|
readme_path.write_text(f"---\n{updated_front}\n---\n{body_text}", encoding="utf-8") |
|
|
logger.info(f"Successfully rewrote {len(results)} URLs in {readme_path}") |
|
|
|
|
|
|
|
|
@app.command() |
|
|
def verify( |
|
|
readme_path: Annotated[ |
|
|
Path, |
|
|
typer.Option( |
|
|
"--readme-path", |
|
|
"-r", |
|
|
help="Path to the README file with YAML front matter.", |
|
|
exists=True, |
|
|
resolve_path=True, |
|
|
dir_okay=False, |
|
|
), |
|
|
] = Path("README.md"), |
|
|
max_concurrency: Annotated[ |
|
|
int, |
|
|
typer.Option( |
|
|
"--max-concurrency", |
|
|
"-c", |
|
|
min=1, |
|
|
show_default=True, |
|
|
help="Maximum concurrent verification requests.", |
|
|
), |
|
|
] = 8, |
|
|
timeout: Annotated[ |
|
|
float, |
|
|
typer.Option( |
|
|
"--timeout", |
|
|
min=1.0, |
|
|
show_default=True, |
|
|
help="Request timeout in seconds.", |
|
|
), |
|
|
] = 60.0, |
|
|
) -> None: |
|
|
front_matter_text, _ = read_front_matter(readme_path) |
|
|
entries = load_data_file_entries(front_matter_text) |
|
|
if not entries: |
|
|
logger.warning("No data_files entries found in README front matter.") |
|
|
raise typer.Exit(code=0) |
|
|
|
|
|
logger.info(f"Verifying {len(entries)} URLs from {readme_path}.") |
|
|
results = asyncio.run( |
|
|
perform_verification( |
|
|
entries=entries, |
|
|
max_concurrency=max_concurrency, |
|
|
timeout=timeout, |
|
|
) |
|
|
) |
|
|
|
|
|
table = Table(title="URL verification results") |
|
|
table.add_column("Config") |
|
|
table.add_column("Status") |
|
|
table.add_column("Detail", overflow="fold") |
|
|
failed = 0 |
|
|
for result in sorted(results, key=lambda r: (not r.ok, r.order)): |
|
|
status_text = str(result.status) if result.status is not None else "-" |
|
|
if result.ok: |
|
|
table.add_row( |
|
|
result.entry.config_name, |
|
|
f"[green]{status_text}[/green]", |
|
|
result.entry.url, |
|
|
) |
|
|
else: |
|
|
failed += 1 |
|
|
detail = result.detail or result.entry.url |
|
|
table.add_row( |
|
|
result.entry.config_name, |
|
|
f"[red]{status_text}[/red]", |
|
|
f"{result.entry.url}\n{detail}", |
|
|
) |
|
|
|
|
|
console.print(table) |
|
|
if failed: |
|
|
logger.error(f"{failed} URLs failed verification.") |
|
|
raise typer.Exit(code=1) |
|
|
logger.info("All URLs verified successfully.") |
|
|
|
|
|
|
|
|
@app.command() |
|
|
def checksum( |
|
|
readme_path: Annotated[ |
|
|
Path, |
|
|
typer.Option( |
|
|
"--readme-path", |
|
|
"-r", |
|
|
help="Path to the README file with YAML front matter.", |
|
|
exists=True, |
|
|
resolve_path=True, |
|
|
dir_okay=False, |
|
|
), |
|
|
] = Path("README.md"), |
|
|
checksum_dir: Annotated[ |
|
|
Path, |
|
|
typer.Option( |
|
|
"--checksum-dir", |
|
|
"-d", |
|
|
help="Directory containing SHA256 checksum files (one per config).", |
|
|
resolve_path=True, |
|
|
), |
|
|
] = Path("checksums"), |
|
|
max_concurrency: Annotated[ |
|
|
int, |
|
|
typer.Option( |
|
|
"--max-concurrency", |
|
|
"-c", |
|
|
min=1, |
|
|
show_default=True, |
|
|
help="Maximum concurrent checksum operations.", |
|
|
), |
|
|
] = 8, |
|
|
timeout: Annotated[ |
|
|
float, |
|
|
typer.Option( |
|
|
"--timeout", |
|
|
min=1.0, |
|
|
show_default=True, |
|
|
help="Request timeout in seconds.", |
|
|
), |
|
|
] = 300.0, |
|
|
generate: Annotated[ |
|
|
bool, |
|
|
typer.Option( |
|
|
"--generate", |
|
|
help="Generate/update checksum files with current remote file checksums.", |
|
|
), |
|
|
] = False, |
|
|
update_mismatches: Annotated[ |
|
|
bool, |
|
|
typer.Option( |
|
|
"--update-mismatches", |
|
|
help="Update checksum files with new hashes for files that don't match.", |
|
|
), |
|
|
] = False, |
|
|
) -> None: |
|
|
"""Verify file integrity using SHA256 checksums.""" |
|
|
front_matter_text, _ = read_front_matter(readme_path) |
|
|
entries = load_data_file_entries(front_matter_text) |
|
|
if not entries: |
|
|
logger.warning("No data_files entries found in README front matter.") |
|
|
raise typer.Exit(code=0) |
|
|
|
|
|
logger.info(f"Computing checksums for {len(entries)} files from {readme_path}.") |
|
|
|
|
|
|
|
|
config_names = sorted(set(entry.config_name for entry in entries)) |
|
|
|
|
|
|
|
|
expected_hashes_by_config: Dict[str, Dict[str, str]] = {} |
|
|
if not generate: |
|
|
for config_name in config_names: |
|
|
expected_hashes_by_config[config_name] = load_checksums_for_config( |
|
|
checksum_dir, config_name |
|
|
) |
|
|
|
|
|
|
|
|
results = asyncio.run( |
|
|
perform_checksum_verification( |
|
|
entries=entries, |
|
|
expected_hashes_by_config=expected_hashes_by_config, |
|
|
max_concurrency=max_concurrency, |
|
|
timeout=timeout, |
|
|
) |
|
|
) |
|
|
|
|
|
|
|
|
table = Table(title="Checksum verification results") |
|
|
table.add_column("Config") |
|
|
table.add_column("Filename") |
|
|
table.add_column("Status") |
|
|
table.add_column("SHA256", overflow="fold") |
|
|
|
|
|
failed = 0 |
|
|
mismatched = 0 |
|
|
|
|
|
new_hashes_by_config: Dict[str, Dict[str, str]] = {} |
|
|
|
|
|
for result in results: |
|
|
if not result.success: |
|
|
failed += 1 |
|
|
error_msg = str(result.error) if result.error else "Unknown error" |
|
|
table.add_row( |
|
|
result.entry.config_name, |
|
|
result.filename, |
|
|
"[red]ERROR[/red]", |
|
|
error_msg, |
|
|
) |
|
|
elif generate: |
|
|
|
|
|
if result.entry.config_name not in new_hashes_by_config: |
|
|
new_hashes_by_config[result.entry.config_name] = {} |
|
|
new_hashes_by_config[result.entry.config_name][result.filename] = ( |
|
|
result.computed_hash or "" |
|
|
) |
|
|
table.add_row( |
|
|
result.entry.config_name, |
|
|
result.filename, |
|
|
"[cyan]COMPUTED[/cyan]", |
|
|
result.computed_hash or "-", |
|
|
) |
|
|
elif result.expected_hash is None: |
|
|
|
|
|
table.add_row( |
|
|
result.entry.config_name, |
|
|
result.filename, |
|
|
"[yellow]NEW[/yellow]", |
|
|
result.computed_hash or "-", |
|
|
) |
|
|
if update_mismatches and result.computed_hash: |
|
|
if result.entry.config_name not in new_hashes_by_config: |
|
|
new_hashes_by_config[result.entry.config_name] = {} |
|
|
new_hashes_by_config[result.entry.config_name][ |
|
|
result.filename |
|
|
] = result.computed_hash |
|
|
elif result.matches: |
|
|
|
|
|
table.add_row( |
|
|
result.entry.config_name, |
|
|
result.filename, |
|
|
"[green]OK[/green]", |
|
|
result.computed_hash or "-", |
|
|
) |
|
|
else: |
|
|
|
|
|
mismatched += 1 |
|
|
table.add_row( |
|
|
result.entry.config_name, |
|
|
result.filename, |
|
|
"[red]MISMATCH[/red]", |
|
|
f"Expected: {result.expected_hash}\nComputed: {result.computed_hash}", |
|
|
) |
|
|
if update_mismatches and result.computed_hash: |
|
|
if result.entry.config_name not in new_hashes_by_config: |
|
|
new_hashes_by_config[result.entry.config_name] = {} |
|
|
new_hashes_by_config[result.entry.config_name][ |
|
|
result.filename |
|
|
] = result.computed_hash |
|
|
|
|
|
console.print(table) |
|
|
|
|
|
|
|
|
if generate: |
|
|
total_saved = 0 |
|
|
for config_name, file_hashes in new_hashes_by_config.items(): |
|
|
save_checksums_for_config(checksum_dir, config_name, file_hashes) |
|
|
total_saved += len(file_hashes) |
|
|
logger.info( |
|
|
f"Generated {len(new_hashes_by_config)} checksum files with {total_saved} total checksums." |
|
|
) |
|
|
elif update_mismatches and new_hashes_by_config: |
|
|
total_updated = 0 |
|
|
for config_name, new_file_hashes in new_hashes_by_config.items(): |
|
|
|
|
|
all_hashes = load_checksums_for_config(checksum_dir, config_name) |
|
|
all_hashes.update(new_file_hashes) |
|
|
save_checksums_for_config(checksum_dir, config_name, all_hashes) |
|
|
total_updated += len(new_file_hashes) |
|
|
logger.warning( |
|
|
f"Updated {total_updated} checksums across {len(new_hashes_by_config)} checksum files." |
|
|
) |
|
|
|
|
|
|
|
|
if failed: |
|
|
logger.error(f"{failed} files failed checksum computation.") |
|
|
raise typer.Exit(code=1) |
|
|
if mismatched and not update_mismatches: |
|
|
logger.error(f"{mismatched} files have checksum mismatches.") |
|
|
logger.warning("Use --update-mismatches to update the checksum files.") |
|
|
raise typer.Exit(code=1) |
|
|
if not generate and not mismatched and not failed: |
|
|
logger.info("All checksums verified successfully.") |
|
|
|
|
|
|
|
|
def clean_arrow_type(type_str: str) -> str: |
|
|
"""Clean up Arrow type string for display.""" |
|
|
|
|
|
|
|
|
cleaned = re.sub(r"list<element:\s*", "list<", type_str) |
|
|
return cleaned |
|
|
|
|
|
|
|
|
def escape_latex(text: str) -> str: |
|
|
"""Escape special LaTeX characters.""" |
|
|
|
|
|
replacements = { |
|
|
"\\": r"\\", |
|
|
"&": r"\&", |
|
|
"%": r"\%", |
|
|
"$": r"\$", |
|
|
"#": r"\#", |
|
|
"_": r"\_", |
|
|
"{": r"\{", |
|
|
"}": r"\}", |
|
|
"~": r"\textasciitilde{}", |
|
|
"^": r"\^{}", |
|
|
} |
|
|
for char, escaped in replacements.items(): |
|
|
text = text.replace(char, escaped) |
|
|
return text |
|
|
|
|
|
|
|
|
async def inspect_file_schema( |
|
|
entry: DataFileEntry, |
|
|
session: aiohttp.ClientSession, |
|
|
semaphore: asyncio.Semaphore, |
|
|
progress: Progress, |
|
|
) -> SchemaInfo: |
|
|
"""Download file and extract schema information using PyArrow.""" |
|
|
_, remote_path = entry.parsed() |
|
|
filename = Path(remote_path).name or remote_path |
|
|
terse_name = (filename[:32] + "…") if len(filename) > 33 else filename |
|
|
description = f"{entry.config_name}: {terse_name}" |
|
|
|
|
|
async with semaphore: |
|
|
task_id: int | None = None |
|
|
temp_file = None |
|
|
try: |
|
|
task_id = progress.add_task(description, total=0, start=False) |
|
|
progress.start_task(task_id) |
|
|
|
|
|
|
|
|
async with session.get(entry.url) as response: |
|
|
response.raise_for_status() |
|
|
total_bytes = response.content_length or 0 |
|
|
|
|
|
if total_bytes: |
|
|
progress.update(task_id, total=total_bytes) |
|
|
|
|
|
with tempfile.NamedTemporaryFile(delete=False, suffix=".parquet") as f: |
|
|
temp_file = Path(f.name) |
|
|
async for chunk in response.content.iter_chunked(1 << 17): |
|
|
f.write(chunk) |
|
|
progress.update(task_id, advance=len(chunk)) |
|
|
|
|
|
|
|
|
parquet_file = pq.read_table(temp_file) |
|
|
schema = parquet_file.schema |
|
|
|
|
|
columns = [ |
|
|
(field.name, clean_arrow_type(str(field.type))) for field in schema |
|
|
] |
|
|
num_rows = len(parquet_file) |
|
|
|
|
|
return SchemaInfo( |
|
|
config_name=entry.config_name, |
|
|
filename=filename, |
|
|
columns=columns, |
|
|
num_rows=num_rows, |
|
|
success=True, |
|
|
) |
|
|
|
|
|
except Exception as exc: |
|
|
return SchemaInfo( |
|
|
config_name=entry.config_name, |
|
|
filename=filename, |
|
|
columns=[], |
|
|
num_rows=None, |
|
|
success=False, |
|
|
error=exc, |
|
|
) |
|
|
finally: |
|
|
if task_id is not None: |
|
|
progress.remove_task(task_id) |
|
|
|
|
|
if temp_file and temp_file.exists(): |
|
|
temp_file.unlink() |
|
|
|
|
|
|
|
|
async def perform_schema_inspection( |
|
|
entries: Sequence[DataFileEntry], |
|
|
max_concurrency: int, |
|
|
timeout: float, |
|
|
) -> List[SchemaInfo]: |
|
|
"""Download first file from each config and inspect schema.""" |
|
|
if not entries: |
|
|
return [] |
|
|
|
|
|
semaphore = asyncio.Semaphore(max_concurrency) |
|
|
timeout_cfg = aiohttp.ClientTimeout(total=timeout) |
|
|
results: List[SchemaInfo] = [] |
|
|
progress = Progress( |
|
|
TextColumn("{task.description}"), |
|
|
BarColumn(bar_width=None), |
|
|
DownloadColumn(), |
|
|
TransferSpeedColumn(), |
|
|
TimeElapsedColumn(), |
|
|
console=console, |
|
|
) |
|
|
|
|
|
async with aiohttp.ClientSession(timeout=timeout_cfg) as session: |
|
|
with progress: |
|
|
tasks = [ |
|
|
asyncio.create_task( |
|
|
inspect_file_schema(entry, session, semaphore, progress) |
|
|
) |
|
|
for entry in entries |
|
|
] |
|
|
for future in asyncio.as_completed(tasks): |
|
|
result = await future |
|
|
results.append(result) |
|
|
|
|
|
|
|
|
results.sort(key=lambda r: r.config_name) |
|
|
return results |
|
|
|
|
|
|
|
|
@app.command() |
|
|
def schema( |
|
|
readme_path: Annotated[ |
|
|
Path, |
|
|
typer.Option( |
|
|
"--readme-path", |
|
|
"-r", |
|
|
help="Path to the README file with YAML front matter.", |
|
|
exists=True, |
|
|
resolve_path=True, |
|
|
dir_okay=False, |
|
|
), |
|
|
] = Path("README.md"), |
|
|
output_file: Annotated[ |
|
|
Path | None, |
|
|
typer.Option( |
|
|
"--output-file", |
|
|
"-o", |
|
|
help="Write schema information to plain text file.", |
|
|
resolve_path=True, |
|
|
dir_okay=False, |
|
|
), |
|
|
] = None, |
|
|
latex: Annotated[ |
|
|
bool, |
|
|
typer.Option( |
|
|
"--latex", |
|
|
help="Output schema in LaTeX format (requires --output-file).", |
|
|
), |
|
|
] = False, |
|
|
max_concurrency: Annotated[ |
|
|
int, |
|
|
typer.Option( |
|
|
"--max-concurrency", |
|
|
"-c", |
|
|
min=1, |
|
|
show_default=True, |
|
|
help="Maximum concurrent downloads.", |
|
|
), |
|
|
] = 4, |
|
|
timeout: Annotated[ |
|
|
float, |
|
|
typer.Option( |
|
|
"--timeout", |
|
|
min=1.0, |
|
|
show_default=True, |
|
|
help="Request timeout in seconds.", |
|
|
), |
|
|
] = 300.0, |
|
|
types: Annotated[bool, typer.Option(help="Add column type info")] = True, |
|
|
) -> None: |
|
|
"""Inspect schema of first file from each dataset config.""" |
|
|
|
|
|
if latex and not output_file: |
|
|
logger.error("--latex requires --output-file to be specified.") |
|
|
raise typer.Exit(code=1) |
|
|
|
|
|
front_matter_text, _ = read_front_matter(readme_path) |
|
|
entries = load_data_file_entries(front_matter_text) |
|
|
if not entries: |
|
|
logger.warning("No data_files entries found in README front matter.") |
|
|
raise typer.Exit(code=0) |
|
|
|
|
|
|
|
|
configs_seen = set() |
|
|
first_files = [] |
|
|
for entry in entries: |
|
|
if entry.config_name not in configs_seen: |
|
|
first_files.append(entry) |
|
|
configs_seen.add(entry.config_name) |
|
|
|
|
|
logger.info(f"Inspecting schema for {len(first_files)} configs.") |
|
|
|
|
|
|
|
|
results = asyncio.run( |
|
|
perform_schema_inspection( |
|
|
entries=first_files, |
|
|
max_concurrency=max_concurrency, |
|
|
timeout=timeout, |
|
|
) |
|
|
) |
|
|
|
|
|
|
|
|
failed = 0 |
|
|
text_output_lines = [] |
|
|
latex_items = [] |
|
|
|
|
|
for result in results: |
|
|
if not result.success: |
|
|
failed += 1 |
|
|
error_msg = str(result.error) if result.error else "Unknown error" |
|
|
logger.error( |
|
|
f"Failed to inspect {result.config_name}/{result.filename}: {error_msg}" |
|
|
) |
|
|
continue |
|
|
|
|
|
|
|
|
if output_file: |
|
|
if latex: |
|
|
|
|
|
escaped_config = escape_latex(result.config_name) |
|
|
column_parts = [] |
|
|
for col_name, col_type in result.columns: |
|
|
escaped_col = escape_latex(col_name) |
|
|
escaped_type = escape_latex(col_type) |
|
|
part = f"\\texttt{{{escaped_col}}}" |
|
|
if types: |
|
|
part += f" (\\texttt{{{escaped_type}}})" |
|
|
column_parts.append(part) |
|
|
columns_str = ", ".join(column_parts) |
|
|
latex_items.append( |
|
|
f"\\item \\textbf{{{escaped_config}}}: {columns_str}" |
|
|
) |
|
|
else: |
|
|
|
|
|
text_output_lines.append(f"# {result.config_name} — {result.filename}") |
|
|
if result.num_rows is not None: |
|
|
text_output_lines.append( |
|
|
f"# {len(result.columns)} columns, {result.num_rows:,} rows" |
|
|
) |
|
|
text_output_lines.append("") |
|
|
for col_name, col_type in result.columns: |
|
|
text_output_lines.append(f"{col_name}: {col_type}") |
|
|
text_output_lines.append("") |
|
|
|
|
|
|
|
|
table = Table(title=f"[bold]{result.config_name}[/bold] — {result.filename}") |
|
|
table.add_column("Column", style="cyan") |
|
|
table.add_column("Type", style="yellow") |
|
|
|
|
|
for col_name, col_type in result.columns: |
|
|
table.add_row(col_name, col_type) |
|
|
|
|
|
console.print(table) |
|
|
if result.num_rows is not None: |
|
|
logger.info( |
|
|
f"{result.config_name}: {len(result.columns)} columns, {result.num_rows:,} rows" |
|
|
) |
|
|
console.print() |
|
|
|
|
|
|
|
|
if output_file: |
|
|
if latex and latex_items: |
|
|
latex_content = ( |
|
|
"\\begin{itemize}\n" + "\n".join(latex_items) + "\n\\end{itemize}" |
|
|
) |
|
|
output_file.write_text(latex_content, encoding="utf-8") |
|
|
logger.info(f"Wrote LaTeX schema to {output_file}") |
|
|
elif text_output_lines: |
|
|
output_file.write_text("\n".join(text_output_lines), encoding="utf-8") |
|
|
logger.info(f"Wrote schema information to {output_file}") |
|
|
|
|
|
if failed: |
|
|
logger.error(f"{failed} configs failed schema inspection.") |
|
|
raise typer.Exit(code=1) |
|
|
logger.info("Schema inspection completed successfully.") |
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
|
app() |
|
|
|