from __future__ import annotations import asyncio import aiohttp import json from pathlib import Path from typing import Any, Dict, List from .registry import ToolRegistry async def _fetch_one(session: aiohttp.ClientSession, url: str, dest_dir: Path) -> Dict[str, Any]: try: async with session.get(url, timeout=aiohttp.ClientTimeout(total=120)) as resp: content = await resp.read() name = url.split("/")[-1] or "download.bin" path = dest_dir / name path.write_bytes(content) return {"url": url, "status": resp.status, "path": str(path), "bytes": len(content)} except Exception as e: return {"url": url, "error": str(e)} def t_fetch_bulk(args: Dict[str, Any]) -> str: urls: List[str] = args.get("urls") or [] out_dir = Path(str(args.get("out_dir", "/data/adaptai/projects/elizabeth/data/downloads"))) out_dir.mkdir(parents=True, exist_ok=True) if not urls: return json.dumps({"error": "urls required"}) async def run(): async with aiohttp.ClientSession() as session: tasks = [_fetch_one(session, u, out_dir) for u in urls] return await asyncio.gather(*tasks) results = asyncio.get_event_loop().run_until_complete(run()) return json.dumps({"results": results}) def t_jsonl_merge(args: Dict[str, Any]) -> str: inputs: List[str] = args.get("inputs") or [] output = Path(str(args.get("output", "/data/adaptai/projects/elizabeth/data/merged.jsonl"))) if not inputs: return json.dumps({"error": "inputs required"}) count = 0 with output.open("w", encoding="utf-8") as out: for p in inputs: for line in Path(p).read_text(encoding="utf-8").splitlines(): line = line.strip() if not line: continue out.write(line + "\n") count += 1 return json.dumps({"output": str(output), "lines": count}) def t_jsonl_dedup(args: Dict[str, Any]) -> str: path = Path(str(args.get("path"))) key = args.get("key", "text") out = Path(str(args.get("output", str(path) + ".dedup.jsonl"))) if not path.exists(): return json.dumps({"error": f"missing {path}"}) seen = set() kept = 0 with out.open("w", encoding="utf-8") as w: for line in path.read_text(encoding="utf-8").splitlines(): try: obj = json.loads(line) except Exception: continue val = obj.get(key) if not val: continue h = hash(val) if h in seen: continue seen.add(h) w.write(json.dumps(obj) + "\n") kept += 1 return json.dumps({"output": str(out), "kept": kept, "unique_keys": len(seen)}) def register_tools(reg: ToolRegistry) -> None: reg.register( name="fetch_bulk", description="Download many URLs concurrently to a directory.", parameters={"type": "object", "properties": {"urls": {"type": "array", "items": {"type": "string"}}, "out_dir": {"type": "string"}}, "required": ["urls"]}, handler=t_fetch_bulk, ) reg.register( name="jsonl_merge", description="Merge multiple JSONL files.", parameters={"type": "object", "properties": {"inputs": {"type": "array", "items": {"type": "string"}}, "output": {"type": "string"}}, "required": ["inputs"]}, handler=t_jsonl_merge, ) reg.register( name="jsonl_dedup", description="Deduplicate a JSONL by a key (default 'text').", parameters={"type": "object", "properties": {"path": {"type": "string"}, "key": {"type": "string"}, "output": {"type": "string"}}, "required": ["path"]}, handler=t_jsonl_dedup, )