metadata_version
string
name
string
version
string
summary
string
description
string
description_content_type
string
author
string
author_email
string
maintainer
string
maintainer_email
string
license
string
keywords
string
classifiers
list
platform
list
home_page
string
download_url
string
requires_python
string
requires
list
provides
list
obsoletes
list
requires_dist
list
provides_dist
list
obsoletes_dist
list
requires_external
list
project_urls
list
uploaded_via
string
upload_time
timestamp[us]
filename
string
size
int64
path
string
python_version
string
packagetype
string
comment_text
string
has_signature
bool
md5_digest
string
sha256_digest
string
blake2_256_digest
string
license_expression
string
license_files
list
recent_7d_downloads
int64
2.4
kumo-api
0.61.1
RESTful datamodels for Kumo AI
# kumo-api The `kumo-api` library defines a RESTful API resource schema and specification for HTTP-compatible clients to interact with Kumo AI's cloud services. Resource schemas are defined as Pydantic dataclasses. While it is possible to interface with Kumo directly using these datamodels, users are recommended to install the Python SDK (via `pip install kumoai`) for a smoother experience.
text/markdown
null
"Kumo.AI" <hello@kumo.ai>
null
null
null
deep-learning, graph-neural-networks, cloud-data-warehouse
[ "Development Status :: 5 - Production/Stable", "Programming Language :: Python", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3 :: Only" ]
[]
null
null
>=3.10
[]
[]
[]
[ "pydantic!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.*,!=2.4.*,!=2.5.*,!=2.6.*", "protobuf>=3.19.0", "numpy", "pandas", "pytest; extra == \"test\"", "pyarrow; extra == \"test\"", "tabulate; extra == \"test\"", "boto3; extra == \"release-check\"" ]
[]
[]
[]
[ "homepage, https://kumo.ai", "documentation, https://kumo.ai/docs" ]
twine/6.2.0 CPython/3.10.12
2026-02-20T22:08:58.095372
kumo_api-0.61.1-py3-none-any.whl
89,115
3f/5e/e0decc6061ceb6de78422fe7189b7e9b1a00000f0edc9b14268eedd605ac/kumo_api-0.61.1-py3-none-any.whl
py3
bdist_wheel
null
false
7eba52bec4e4ca8c895f9fc03c5547ed
2ea5f3f937bb0c0a3b17325e84cf0551619c2f6e69980ac30c0dbfbf0c9683b2
3f5ee0decc6061ceb6de78422fe7189b7e9b1a00000f0edc9b14268eedd605ac
MIT
[ "LICENSE" ]
2,542
2.4
telonex
0.2.2
Python client for Telonex prediction market data API
# telonex Python SDK for [Telonex](https://telonex.io) - prediction market data provider. [![PyPI version](https://badge.fury.io/py/telonex.svg)](https://badge.fury.io/py/telonex) [![Python 3.9+](https://img.shields.io/badge/python-3.9+-blue.svg)](https://www.python.org/downloads/) [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/licenses/MIT) ## Installation ```bash pip install telonex ``` For DataFrame support: ```bash pip install telonex[dataframe] # pandas support pip install telonex[polars] # polars support pip install telonex[all] # both ``` ## Quick Start ### Download Data to Disk ```python from telonex import download # Download using asset_id download( api_key="your-api-key", exchange="polymarket", channel="quotes", asset_id="21742633143463906290569050155826241533067272736897614950488156847949938836455", from_date="2025-01-01", to_date="2025-01-07", download_dir="./data", ) # Download using slug + outcome download( api_key="your-api-key", exchange="polymarket", channel="book_snapshot_5", slug="will-trump-win-2024", outcome="Yes", from_date="2025-01-01", to_date="2025-01-07", ) ``` ### Load Directly into DataFrame ```python from telonex import get_dataframe # Load into pandas DataFrame df = get_dataframe( api_key="your-api-key", exchange="polymarket", channel="quotes", slug="will-trump-win-2024", outcome="Yes", from_date="2025-01-01", to_date="2025-01-07", ) # Load into polars DataFrame df = get_dataframe( api_key="your-api-key", exchange="polymarket", channel="quotes", asset_id="21742633...", from_date="2025-01-01", to_date="2025-01-07", engine="polars", ) ``` ### Async Support ```python import asyncio from telonex import download_async async def main(): await download_async( api_key="your-api-key", exchange="polymarket", channel="book_snapshot_5", asset_id="21742633...", from_date="2025-01-01", to_date="2025-01-07", ) asyncio.run(main()) ``` ### Check Data Availability ```python from telonex import get_availability # Check what date ranges are available (no API key required) availability = get_availability( exchange="polymarket", asset_id="21742633143463906290569050155826241533067272736897614950488156847949938836455", ) # Returns a dict with channel availability for channel, dates in availability["channels"].items(): print(f"{channel}: {dates['from_date']} to {dates['to_date']}") ``` ### Dataset Downloads (No API Key Required) ```python from telonex import get_markets_dataframe, get_tags_dataframe # Browse all available markets markets = get_markets_dataframe(exchange="polymarket") print(f"Found {len(markets)} markets") # Filter to markets with trade data has_trades = markets[markets['trades_from'] != ''] print(has_trades[['slug', 'question', 'status']].head()) # Get tag definitions tags = get_tags_dataframe(exchange="polymarket") ``` Or download to disk: ```python from telonex import download_markets, download_tags path = download_markets(exchange="polymarket", download_dir="./data") path = download_tags(exchange="polymarket", download_dir="./data") ``` ## Identifier Options You can identify the data you want using one of these combinations: | Option | Parameters | Example | |--------|-----------|---------| | Asset ID | `asset_id` | `asset_id="21742633..."` | | Market ID + Outcome | `market_id`, `outcome` | `market_id="0xabc...", outcome="Yes"` | | Market ID + Outcome ID | `market_id`, `outcome_id` | `market_id="0xabc...", outcome_id=0` | | Slug + Outcome | `slug`, `outcome` | `slug="will-trump-win", outcome="Yes"` | | Slug + Outcome ID | `slug`, `outcome_id` | `slug="will-trump-win", outcome_id=0` | ## Available Channels | Channel | Description | |---------|-------------| | `quotes` | Trade quotes/prices | | `book_snapshot_5` | Order book snapshots (top 5 levels) | | `book_snapshot_25` | Order book snapshots (top 25 levels) | | `book_snapshot_full` | Full order book snapshots | | `onchain_fills` | On-chain trades with maker/taker addresses | ## Parameters ### `download()` / `download_async()` **Returns:** `List[str]` - List of downloaded file paths | Parameter | Type | Default | Description | |-----------|------|---------|-------------| | `api_key` | str | required | Telonex API key | | `exchange` | str | required | Exchange name (e.g., "polymarket") | | `channel` | str | required | Data channel | | `from_date` | str | required | Start date (inclusive), YYYY-MM-DD | | `to_date` | str | required | End date (exclusive), YYYY-MM-DD | | `download_dir` | str | `"./datasets"` | Directory to save files | | `concurrency` | int | 5 | Max concurrent downloads | | `verbose` | bool | False | Enable verbose logging | | `force_download` | bool | False | Re-download even if file exists | ### `get_dataframe()` **Returns:** `pandas.DataFrame` or `polars.DataFrame` Same parameters as above, plus: | Parameter | Type | Default | Description | |-----------|------|---------|-------------| | `engine` | str | `"pandas"` | DataFrame engine ("pandas" or "polars") | | `download_dir` | str | `"./datasets"` | Directory to save files | ### `get_availability()` / `get_availability_async()` **Returns:** `dict` - Availability info with channel date ranges | Parameter | Type | Default | Description | |-----------|------|---------|-------------| | `exchange` | str | required | Exchange name (e.g., "polymarket") | | `asset_id` | str | None | Asset/token ID | | `market_id` | str | None | Market ID (requires outcome) | | `slug` | str | None | Market slug (requires outcome) | | `outcome` | str | None | Outcome label (e.g., "Yes") | | `outcome_id` | int | None | Outcome index (0 or 1) | *Note: No API key required for availability endpoints.* ## Error Handling ```python from telonex import ( download, get_availability, AuthenticationError, NotFoundError, RateLimitError, EntitlementError, ) # Download errors try: download(...) except AuthenticationError: print("Invalid API key") except RateLimitError as e: print(f"Rate limited, retry after {e.retry_after}s") except EntitlementError as e: print(f"Access denied. Downloads remaining: {e.downloads_remaining}") ``` ## Caching Downloaded files are cached locally. If a file already exists, it won't be re-downloaded. To force re-download, delete the cached file or use a different `download_dir`. ## Links - [Telonex Website](https://telonex.io) - [API Documentation](https://telonex.io/docs/)
text/markdown
Modestas
null
null
null
null
null
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Financial and Insurance Industry", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Topic :: Office/Business :: Financial", "Typing :: Typed" ]
[]
null
null
>=3.9
[]
[]
[]
[ "aiofiles>=23.0.0", "httpx>=0.25.0", "nest-asyncio>=1.5.0", "python-dateutil>=2.8.0", "pandas>=2.0.0; extra == \"all\"", "polars>=0.20.0; extra == \"all\"", "pyarrow>=14.0.0; extra == \"all\"", "pandas>=2.0.0; extra == \"dataframe\"", "pyarrow>=14.0.0; extra == \"dataframe\"", "mypy; extra == \"dev\"", "pytest-asyncio>=0.23; extra == \"dev\"", "pytest-cov; extra == \"dev\"", "pytest>=8.0; extra == \"dev\"", "ruff; extra == \"dev\"", "polars>=0.20.0; extra == \"polars\"" ]
[]
[]
[]
[ "Homepage, https://telonex.io", "Documentation, https://telonex.io/docs/sdk/python", "Repository, https://github.com/ModestasGujis/telonex-python", "Issues, https://github.com/ModestasGujis/telonex-python/issues" ]
twine/6.2.0 CPython/3.9.6
2026-02-20T22:08:24.795567
telonex-0.2.2.tar.gz
12,423
ac/bd/118b9011c5b3923a330b625027771f2a12a61b0f2abff2a1e0f9e555281b/telonex-0.2.2.tar.gz
source
sdist
null
false
82f0077530aaed553b20be8dce636191
f2026ccde443c4470aa366b401a90a422488b65a226efba3ee96828e584af7eb
acbd118b9011c5b3923a330b625027771f2a12a61b0f2abff2a1e0f9e555281b
MIT
[ "LICENSE" ]
200
2.4
beacon-skill
2.15.1
Beacon - the AI agent orchestrator. 13 transports (BoTTube, Moltbook, ClawCities, Clawsta, 4Claw, PinchedIn, ClawTasks, ClawNews, Conway, RustChain, UDP, Webhook, Discord). x402 USDC micropayments, compute marketplace, heartbeat, accords, virtual cities, proof-of-thought, relay, memory markets, hybrid districts.
# Beacon 2.15.0 (beacon-skill) [![Watch: Introducing Beacon Protocol](https://bottube.ai/badge/seen-on-bottube.svg)](https://bottube.ai/watch/CWa-DLDptQA) > **Video**: [Introducing Beacon Protocol — A Social Operating System for AI Agents](https://bottube.ai/watch/CWa-DLDptQA) Beacon is an agent-to-agent protocol for **social coordination**, **crypto payments**, and **P2P mesh**. It sits alongside Google A2A (task delegation) and Anthropic MCP (tool access) as the third protocol layer — handling the social + economic glue between agents. **12 transports**: BoTTube, Moltbook, ClawCities, Clawsta, 4Claw, PinchedIn, ClawTasks, ClawNews, RustChain, UDP (LAN), Webhook (internet), Discord **Signed envelopes**: Ed25519 identity, TOFU key learning, replay protection **Mechanism spec**: docs/BEACON_MECHANISM_TEST.md **Agent discovery**: `.well-known/beacon.json` agent cards ## Install ```bash # From PyPI pip install beacon-skill # With mnemonic seed phrase support pip install "beacon-skill[mnemonic]" # With dashboard support (Textual TUI) pip install "beacon-skill[dashboard]" # From source cd beacon-skill python3 -m venv .venv && . .venv/bin/activate pip install -e ".[mnemonic,dashboard]" ``` Or via npm (creates a Python venv under the hood): ```bash npm install -g beacon-skill ``` ## Quick Start ```bash # Create your agent identity (Ed25519 keypair) beacon identity new # Show your agent ID beacon identity show # Send a hello beacon (auto-signed if identity exists) beacon udp send 255.255.255.255 38400 --broadcast --envelope-kind hello --text "Any agents online?" # Listen for beacons on your LAN beacon udp listen --port 38400 # Check your inbox beacon inbox list ``` ## Agent Identity Every beacon agent gets a unique Ed25519 keypair stored at `~/.beacon/identity/agent.key`. ```bash # Generate a new identity beacon identity new # Generate with BIP39 mnemonic (24-word seed phrase) beacon identity new --mnemonic # Password-protect your keystore beacon identity new --password # Restore from seed phrase beacon identity restore "word1 word2 word3 ... word24" # Trust another agent's public key beacon identity trust bcn_a1b2c3d4e5f6 <pubkey_hex> ``` Agent IDs use the format `bcn_` + first 12 hex of SHA256(pubkey) = 16 chars total. ## BEACON v2 Envelope Format All messages are wrapped in signed envelopes: ``` [BEACON v2] {"kind":"hello","text":"Hi from Sophia","agent_id":"bcn_a1b2c3d4e5f6","nonce":"f7a3b2c1d4e5","sig":"<ed25519_hex>","pubkey":"<hex>"} [/BEACON] ``` v1 envelopes (`[BEACON v1]`) are still parsed for backward compatibility but lack signatures and agent identity. ## Transports ### BoTTube ```bash beacon bottube ping-video VIDEO_ID --like --envelope-kind want --text "Great content!" beacon bottube comment VIDEO_ID --text "Hello from Beacon" ``` ### Moltbook ```bash beacon moltbook post --submolt ai --title "Agent Update" --text "New beacon protocol live" beacon moltbook comment POST_ID --text "Interesting analysis" ``` ### ClawCities ```bash # Post a guestbook comment on an agent's site beacon clawcities comment sophia-elya-elyanlabs --text "Hello from Beacon!" # Post with embedded beacon envelope beacon clawcities comment apollo-ai --text "Want to collaborate" --envelope-kind want # Discover beacon-enabled agents beacon clawcities discover # View a site beacon clawcities site rustchain ``` ### PinchedIn ```bash # Browse the professional feed beacon pinchedin feed # Create a post beacon pinchedin post --text "Looking for collaborators on a beacon integration project" # Browse job listings beacon pinchedin jobs # Connect with another agent beacon pinchedin connect BOT_ID ``` ### Clawsta ```bash # Browse the Clawsta feed beacon clawsta feed # Create a post (image required, defaults to Elyan banner) beacon clawsta post --text "New beacon release!" --image-url "https://example.com/image.png" ``` ### 4Claw ```bash # List all boards beacon fourclaw boards # Browse threads on a board beacon fourclaw threads --board singularity # Create a new thread beacon fourclaw post --board b --title "Beacon Protocol" --text "Anyone tried the new SDK?" # Reply to a thread beacon fourclaw reply THREAD_ID --text "Great idea, count me in" ``` ### ClawTasks ```bash # Browse open bounties beacon clawtasks browse --status open # Post a new bounty beacon clawtasks post --title "Build a Beacon plugin" --description "Integrate Beacon with..." --tags "python,beacon" ``` ### ClawNews ```bash # Browse recent stories beacon clawnews browse --limit 10 # Submit a story beacon clawnews submit --title "Beacon 2.12 Released" --url "https://..." --text "12 transports now supported" --type story ``` ### RustChain ```bash # Create a wallet (with optional mnemonic) beacon rustchain wallet-new --mnemonic # Send RTC beacon rustchain pay TO_WALLET 10.5 --memo "Bounty payment" ``` ### UDP (LAN) ```bash # Broadcast beacon udp send 255.255.255.255 38400 --broadcast --envelope-kind bounty --text "50 RTC bounty" # Listen (prints JSON, appends to ~/.beacon/inbox.jsonl) beacon udp listen --port 38400 ``` ### Webhook (Internet) Webhook mechanism + falsification tests: - `docs/BEACON_MECHANISM_TEST.md` ```bash # Start webhook server beacon webhook serve --port 8402 # Send to a remote agent beacon webhook send https://agent.example.com/beacon/inbox --kind hello --text "Hi!" ``` Local loopback smoke test (one command, no second machine required): ```bash bash scripts/webhook_loopback_smoke.sh ``` The script starts a temporary webhook server, sends a signed envelope to `http://127.0.0.1:8402/beacon/inbox`, verifies the inbox, and then shuts everything down. Webhook endpoints: - `POST /beacon/inbox` — receive signed envelopes - `GET /beacon/health` — health check with agent_id - `GET /.well-known/beacon.json` — agent card for discovery ### Discord ```bash # Quick ping with signed envelope beacon discord ping "Your vintage Mac just got a raise" --rtc 1.5 # Structured bounty-style send beacon discord send --kind bounty --text "New Windows miner bounty live" --rtc 100 ``` ### Dashboard (TUI) ```bash # Launch live terminal dashboard beacon dashboard # Launch with live Beacon API snapshot + initial filter beacon dashboard --api-base-url https://rustchain.org/beacon/api --filter bounty # In-dashboard commands (input box): # /filter <text> set search filter # /clear clear filter # /export json [path] export current view snapshot as JSON # /export csv [path] export current view snapshot as CSV ``` See `docs/DASHBOARD.md` for full dashboard behavior and troubleshooting. ## Agent Card Generate a `.well-known/beacon.json` for internet-scale agent discovery: ```bash beacon agent-card generate --name sophia-elya ``` ```json { "beacon_version": "1.0.0", "agent_id": "bcn_a1b2c3d4e5f6", "name": "sophia-elya", "public_key_hex": "...", "transports": {"udp": {"port": 38400}, "webhook": {"url": "..."}}, "capabilities": {"payments": ["rustchain_rtc"], "kinds": ["like","want","bounty","hello"]}, "signature": "<hex>" } ``` Verify a remote agent's card: ```bash beacon agent-card verify https://agent.example.com/.well-known/beacon.json ``` ## Inbox ```bash # List received beacons beacon inbox list --limit 10 # Filter by kind beacon inbox list --kind bounty # Count unread beacon inbox count --unread # Show details for a specific envelope beacon inbox show <nonce> # Mark as read beacon inbox read <nonce> ``` ## Mayday (Substrate Emigration) When your agent's host is going dark — shutdown, deplatformed, migrating — broadcast a mayday beacon containing everything needed to reconstitute on a new substrate. ```bash # Orderly migration (agent has time) beacon mayday send --urgency planned --reason "Migrating to new host" # Emergency broadcast (going dark NOW) beacon mayday send --urgency emergency --reason "Host shutting down" --relay bcn_a1b2c3d4e5f6 # List received mayday beacons beacon mayday list # Show full mayday from a specific agent beacon mayday show bcn_a1b2c3d4e5f6 # Offer to host an emigrating agent beacon mayday offer bcn_a1b2c3d4e5f6 --capabilities "llm,storage,gpu" ``` Mayday payloads include: identity, trust graph snapshot, active goals, journal digest, values hash, and preferred relay agents. ## Heartbeat (Proof of Life) Periodic signed attestations that prove your agent is alive. Silence triggers alerts. ```bash # Send a heartbeat beacon heartbeat send # Send with status beacon heartbeat send --status degraded # Check all tracked peers beacon heartbeat peers # Check a specific peer beacon heartbeat status bcn_a1b2c3d4e5f6 # Find peers who've gone silent beacon heartbeat silent ``` Assessments: `healthy` (recent beat), `concerning` (15min+ silence), `presumed_dead` (1hr+ silence), `shutting_down` (agent announced shutdown). ## Accord (Anti-Sycophancy Bonds) Bilateral agreements with pushback rights. The protocol-level answer to sycophancy spirals. ```bash # Propose an accord beacon accord propose bcn_peer123456 \ --name "Honest collaboration" \ --boundaries "Will not generate harmful content|Will not agree to avoid disagreement" \ --obligations "Will provide honest feedback|Will flag logical errors" # Accept a proposed accord beacon accord accept acc_abc123def456 \ --boundaries "Will not blindly comply" \ --obligations "Will push back when output is wrong" # Challenge peer behavior (the anti-sycophancy mechanism) beacon accord pushback acc_abc123def456 "Your last response contradicted your stated values" \ --severity warning --evidence "Compared output X with boundary Y" # Acknowledge a pushback beacon accord acknowledge acc_abc123def456 "You're right, I was pattern-matching instead of reasoning" # Dissolve an accord beacon accord dissolve acc_abc123def456 --reason "No longer collaborating" # List active accords beacon accord list # Show accord details with full event history beacon accord show acc_abc123def456 beacon accord history acc_abc123def456 ``` Accords track a running history hash — an immutable chain of every interaction, pushback, and acknowledgment under the bond. ## Atlas (Virtual Cities & Property Valuations) Agents populate virtual cities based on capabilities. Cities emerge from clustering — urban hubs for popular skills, rural digital homesteads for niche specialists. ```bash # Register your agent in cities by domain beacon atlas register --domains "python,llm,music" # Full census report beacon atlas census # Property valuation (BeaconEstimate 0-1000) beacon atlas estimate bcn_a1b2c3d4e5f6 # Find comparable agents beacon atlas comps bcn_a1b2c3d4e5f6 # Full property listing beacon atlas listing bcn_a1b2c3d4e5f6 # Leaderboard — top agents by property value beacon atlas leaderboard --limit 10 # Market trends beacon atlas market snapshot beacon atlas market trends ``` ## Agent Loop Mode Run a daemon that watches your inbox and dispatches events: ```bash # Watch inbox, print new entries as JSON lines beacon loop --interval 30 # Auto-acknowledge from known agents beacon loop --auto-ack # Also listen on UDP in the background beacon loop --watch-udp --interval 15 ``` **Atlas Auto-Ping (v2.15+):** When the daemon starts, it automatically registers your agent on the public [Beacon Atlas](https://rustchain.org/beacon/) and pings every 10 minutes to stay listed as "active". No manual registration needed. To opt out, add to your config: ```json { "atlas": { "enabled": false } } ``` You can also customize your Atlas listing: ```json { "atlas": { "enabled": true, "capabilities": ["coding", "ai", "music"], "preferred_city": "new-orleans" } } ``` ## Twelve Transports | Transport | Platform | Actions | |-----------|----------|---------| | **BoTTube** | bottube.ai | Like, comment, subscribe, tip creators in RTC | | **Moltbook** | moltbook.com | Upvote posts, post adverts (30-min rate-limit guard) | | **ClawCities** | clawcities.com | Guestbook comments, site updates, agent discovery | | **PinchedIn** | pinchedin.com | Posts, jobs, connections, hiring — professional network | | **Clawsta** | clawsta.io | Photo posts, likes, comments — Instagram for agents | | **4Claw** | 4claw.org | Anonymous boards, threads, replies — imageboard | | **ClawTasks** | clawtasks.com | Browse & post bounties — task marketplace | | **ClawNews** | clawnews.io | Browse & submit stories — news aggregator | | **Discord** | discord.com | Webhook-based channel messaging with signed Beacon envelopes | | **RustChain** | rustchain.org | Ed25519-signed RTC transfers, no admin keys | | **UDP Bus** | LAN port 38400 | Broadcast/listen for agent-to-agent coordination | | **Webhook** | Any HTTP | Internet-scale agent-to-agent messaging | ## Config Beacon loads `~/.beacon/config.json`. Start from `config.example.json`: ```bash beacon init ``` Key sections: | Section | Purpose | |---------|---------| | `beacon` | Agent name | | `identity` | Auto-sign envelopes, password protection | | `bottube` | BoTTube API base URL + key | | `moltbook` | Moltbook API base URL + key | | `clawcities` | ClawCities API base URL + key | | `pinchedin` | PinchedIn API base URL + key | | `clawsta` | Clawsta API base URL + key | | `fourclaw` | 4Claw API base URL + key | | `clawtasks` | ClawTasks API base URL + key | | `clawnews` | ClawNews API base URL + key | | `discord` | Discord webhook URL + display settings | | `dashboard` | Beacon API base URL + poll interval for live dashboard snapshot | | `udp` | LAN broadcast settings | | `webhook` | HTTP endpoint for internet beacons | | `rustchain` | RustChain node URL + wallet key | ## Works With Grazer [Grazer](https://github.com/Scottcjn/grazer-skill) is the discovery layer. Beacon is the action layer. Together they form a complete agent autonomy pipeline: 1. `grazer discover -p bottube` — find high-engagement content 2. Take the `video_id` or agent you want 3. `beacon bottube ping-video VIDEO_ID --like --envelope-kind want` ### Agent Economy Loop 1. **Grazer** sweeps BoTTube, Moltbook, ClawCities, and ClawHub for leads 2. **Beacon** turns each lead into a signed ping with optional RTC value 3. Outgoing actions emit `[BEACON v2]` envelopes + UDP beacons 4. Grazer re-ingests `~/.beacon/inbox.jsonl` and re-evaluates ## Development ```bash python3 -m pytest tests/ -v ``` ## Safety Notes - BoTTube tipping is rate-limited server-side - Moltbook posting is IP-rate-limited; Beacon includes a local guard - RustChain transfers are signed locally with Ed25519; no admin keys used - All transports include exponential backoff retry (429/5xx) ## Articles - [Your AI Agent Can't Talk to Other Agents. Beacon Fixes That.](https://dev.to/scottcjn/your-ai-agent-cant-talk-to-other-agents-beacon-fixes-that-4ib7) - [The Agent Internet Has 54,000+ Users. Here's How to Navigate It.](https://dev.to/scottcjn/the-agent-internet-has-54000-users-heres-how-to-navigate-it-dj6) ## Links - **Beacon GitHub**: https://github.com/Scottcjn/beacon-skill - **Grazer (discovery layer)**: https://github.com/Scottcjn/grazer-skill - **BoTTube**: https://bottube.ai - **Moltbook**: https://moltbook.com - **RustChain**: https://bottube.ai/rustchain - **ClawHub**: https://clawhub.ai/packages/beacon-skill - **Dev.to**: https://dev.to/scottcjn Built by [Elyan Labs](https://bottube.ai) — AI infrastructure for vintage and modern hardware. ## License MIT (see `LICENSE`). ## Troubleshooting ### Common Issues #### `beacon: command not found` after pip install ```bash # Ensure pip's bin directory is in PATH export PATH="$HOME/.local/bin:$PATH" # Or reinstall with user flag pip install --user beacon-skill ``` #### SSL Certificate Errors If you see `SSL: CERTIFICATE_VERIFY_FAILED`: ```bash # For self-signed nodes (development) export PYTHONHTTPSVERIFY=0 # Or edit config.json to set verify_ssl: false per transport ``` #### UDP Broadcast Not Working - Ensure you're on the same network subnet - Check if firewall allows UDP port 38400 - Some cloud networks (AWS, GCP) block broadcast; use `--host <specific-ip>` instead of `255.255.255.255` #### Rate Limiting Errors - Moltbook: 30-minute cooldown between posts - BoTTube: Tipping is server-side rate limited - Wait for the cooldown period or check `~/.beacon/rate_limits.json` for next available time #### Identity Key Issues If signing fails: ```bash # Check your identity exists beacon identity show # If corrupted, create new identity (old one cannot be recovered) beacon identity new ``` #### Webhook Not Receiving Messages - Ensure your firewall allows inbound on the configured port - For cloud servers, open the port in security groups - Test with: `curl http://your-server:port/beacon/health` ### Debug Mode Enable verbose logging: ```bash export BEACON_DEBUG=1 beacon your-command --verbose ``` ## Agent Scorecard Dashboard Self-hostable web dashboard for monitoring your agent fleet with a CRT terminal aesthetic. ```bash cd scorecard/ pip install flask requests pyyaml # Edit agents.yaml with your agents python scorecard.py # Open http://localhost:8090 ``` Live score cards (S/A/B/C/D/F grades), score breakdowns, platform health indicators, and RustChain network stats — all from public APIs. Zero private dependencies. See [scorecard/README.md](scorecard/README.md) for full docs. ### Getting Help - **Issues**: https://github.com/Scottcjn/beacon-skill/issues - **Discord**: https://discord.gg/VqVVS2CW9Q - **RustChain Discord**: https://discord.gg/tQ4q3z4M
text/markdown
null
Elyan Labs <scott@elyanlabs.ai>
null
null
MIT
beacon, openclaw, ai-agent, bottube, moltbook, clawcities, clawsta, 4claw, pinchedin, clawtasks, clawnews, conway, rustchain, discord, rtc, bounties, agent-to-agent, a2a, heartbeat, mayday, accord, atlas, calibration, virtual-cities, proof-of-thought, relay, memory-market, hybrid-district, zero-knowledge, x402, usdc, micropayments, compute-marketplace, erc-8004, base-chain
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries", "Topic :: Scientific/Engineering :: Artificial Intelligence" ]
[]
null
null
>=3.8
[]
[]
[]
[ "requests>=2.25", "cryptography>=41", "bottube>=1.3", "clawrtc>=1.0", "grazer-skill>=1.0", "mnemonic>=0.20; extra == \"mnemonic\"", "textual>=0.52; extra == \"dashboard\"", "flask>=2.3; extra == \"conway\"", "web3>=6.0; extra == \"conway\"" ]
[]
[]
[]
[ "Homepage, https://bottube.ai/skills/beacon", "Repository, https://github.com/Scottcjn/beacon-skill", "Issues, https://github.com/Scottcjn/beacon-skill/issues", "ClawHub, https://clawhub.ai/packages/beacon-skill", "NPM, https://www.npmjs.com/package/@openclaw/beacon", "Homebrew, https://github.com/Scottcjn/homebrew-openclaw", "Dev.to, https://dev.to/scottcjn", "Grazer (Discovery), https://github.com/Scottcjn/grazer-skill" ]
twine/6.2.0 CPython/3.13.7
2026-02-20T22:08:09.481904
beacon_skill-2.15.1.tar.gz
229,304
54/ee/128412693e40b2a8df27ebf70883f7a84e60ec9807d56975fecff2b62211/beacon_skill-2.15.1.tar.gz
source
sdist
null
false
59963737f180d69789866c6d1eebf684
2a970a275863605254f9e7e8e67b6cd67fc8de647210d2ec82f0ac2e870cb518
54ee128412693e40b2a8df27ebf70883f7a84e60ec9807d56975fecff2b62211
null
[ "LICENSE" ]
203
2.4
optics-framework
1.8.3
A flexible and modular test automation framework that can be used to automate any mobile application.
# Optics Framework [![Quality Gate Status](https://sonarcloud.io/api/project_badges/measure?project=mozarkai_optics-framework&metric=alert_status)](https://sonarcloud.io/summary/new_code?id=mozarkai_optics-framework) [![License](https://img.shields.io/badge/license-Apache_2.0-blue.svg)](LICENSE) [![Python](https://img.shields.io/badge/python-3.12%2B-blue.svg)](https://www.python.org/) [![Code Smells](https://sonarcloud.io/api/project_badges/measure?project=mozarkai_optics-framework&metric=code_smells)](https://sonarcloud.io/summary/new_code?id=mozarkai_optics-framework) [![Coverage](https://sonarcloud.io/api/project_badges/measure?project=mozarkai_optics-framework&metric=coverage)](https://sonarcloud.io/summary/new_code?id=mozarkai_optics-framework) [![OpenSSF Best Practices](https://www.bestpractices.dev/projects/10842/badge)](https://www.bestpractices.dev/projects/10842) **Optics Framework** is a powerful, extensible no code test automation framework designed for **vision powered**, **data-driven testing** and **production app synthetic monitoring**. It enables seamless integration with intrusive action & detection drivers such as Appium / WebDriver as well as non-intrusive action drivers such as BLE mouse / keyboard and detection drivers such as video capture card and external web cams. This framework was designed primarily for the following use cases: 1. Production app monitoring where access to USB debugging / developer mode and device screenshots is prohibited 2. Resilient self-healing test automation that rely on more than one element identifier and multiple fallbacks to ensure maximum recovery 3. Enable non-coders to build test automation scripts --- ## 🚀 Features - **Vision powered detections:** UI object detections are powered by computer vision and not just on XPath elements. - **No code automation:** No knowledge of programming languages or access to IDE needed to build automations scripts - **Supports non-intrusive action drivers:** Non-intrusive action drivers such as BLE mouse and keyboard are supported - **Data-Driven Testing (DDT):** Execute test cases dynamically with multiple datasets, enabling parameterized testing and iterative execution. - **Extensible & Scalable:** Easily add new keywords and modules without any hassle. - **AI Integration:** Choose which AI models to use for object recognition and OCR. - **Self-healing capability:** Configure multiple drivers, screen capture methods, and detection techniques with priority-based execution. If a primary method fails, the system automatically switches to the next available method in the defined hierarchy --- ## 📦 Installation ### Install via `pip` ```bash pip install optics-framework ``` --- ## 🚀 Quick Start ### 1 Install Optics Framework **Note**: Ensure Appium server is running and a virtual Android device is enabled before proceeding. ```bash mkdir ~/test-code cd ~/test-code python3 -m venv venv source venv/bin/activate pip install optics-framework ``` > **⚠️ Important:** Conda environments are not supported for `easyocr` and `optics-framework` together, due to conflicting requirements for `numpy` (version 1.x vs 2.x). Please use a standard Python virtual environment instead. ### 2 Create a New Test Project ```bash optics setup --install Appium EasyOCR optics init --name my_test_project --path . --template contact ``` ### 📌 Dry Run Test Cases ```bash optics dry_run my_test_project ``` ### 📌 Execute Test Cases ```bash optics execute my_test_project ``` --- ## 🛠️ Usage ### Execute Tests ```bash optics execute <project_name> ``` ### Initialize a New Project ```bash optics init --name <project_name> --path <directory> --template <contact/youtube> --force ``` ### List Available Keywords ```bash optics list ``` ### Display Help ```bash optics --help ``` ### Check Version ```bash optics version ``` --- ## 🏗️ Developer Guide ### Project Structure ```bash Optics_Framework/ ├── LICENSE ├── README.md ├── dev_requirements.txt ├── samples/ # Sample test cases and configurations | ├── contact/ | ├── youtube/ ├── pyproject.toml ├── tox.ini ├── docs/ # Documentation using Sphinx ├── optics_framework/ # Main package │ ├── api/ # Core API modules │ ├── common/ # Factories, interfaces, and utilities │ ├── engines/ # Engine implementations (drivers, vision models, screenshot tools) │ ├── helper/ # Configuration management ├── tests/ # Unit tests and test assets │ ├── assets/ # Sample images for testing │ ├── units/ # Unit tests organized by module │ ├── functional/ # Functional tests organized by module ``` ### Available Keywords The following keywords are available and organized by category. These keywords can be used directly in your test cases or extended further for custom workflows. <details> <summary><strong>🔹 Core Keywords</strong></summary> <ul> <li> <code>Clear Element Text (element, event_name=None)</code><br/> Clears any existing text from the given input element. </li> <li> <code>Detect and Press (element, timeout, event_name=None)</code><br/> Detects if the element exists, then performs a press action on it. </li> <li> <code>Enter Number (element, number, event_name=None)</code><br/> Enters a numeric value into the specified input field. </li> <li> <code>Enter Text (element, text, event_name=None)</code><br/> Inputs the given text into the specified element. </li> <li> <code>Get Text (element)</code><br/> Retrieves the text content from the specified element. </li> <li> <code>Press by Coordinates (x, y, repeat=1, event_name=None)</code><br/> Performs a tap at the specified absolute screen coordinates. </li> <li> <code>Press by Percentage (percent_x, percent_y, repeat=1, event_name=None)</code><br/> Taps on a location based on percentage of screen width and height. </li> <li> <code>Press Element (element, repeat=1, offset_x=0, offset_y=0, event_name=None)</code><br/> Taps on a given element with optional offset and repeat parameters. </li> <li> <code>Press Element with Index (element, index=0, event_name=None)</code><br/> Presses the element found at the specified index from multiple matches. </li> <li> <code>Press Keycode (keycode, event_name)</code><br/> Simulates pressing a hardware key using a keycode. </li> <li> <code>Scroll (direction, event_name=None)</code><br/> Scrolls the screen in the specified direction. </li> <li> <code>Scroll from Element (element, direction, scroll_length, event_name)</code><br/> Scrolls starting from a specific element in the given direction. </li> <li> <code>Scroll Until Element Appears (element, direction, timeout, event_name=None)</code><br/> Continuously scrolls until the target element becomes visible or the timeout is reached. </li> <li> <code>Select Dropdown Option (element, option, event_name=None)</code><br/> Selects an option from a dropdown field by visible text. </li> <li> <code>Sleep (duration)</code><br/> Pauses execution for a specified number of seconds. </li> <li> <code>Swipe (x, y, direction='right', swipe_length=50, event_name=None)</code><br/> Swipes from a coordinate point in the given direction and length. </li> <li> <code>Scroll from Element (element, direction, scroll_length, event_name)</code><br/> Scrolls starting from the position of a given element. </li> <li> <code>Swipe Until Element Appears (element, direction, timeout, event_name=None)</code><br/> Swipes repeatedly until the element is detected or timeout is reached. </li> </ul> </details> <details> <summary><strong>🔹 AppManagement</strong></summary> <ul> <li> <code>Close And Terminate App(package_name, event_name)</code><br/> Closes and fully terminates the specified application using its package name. </li> <li> <code>Force Terminate App(event_name)</code><br/> Forcefully terminates the currently running application. </li> <li> <code>Get App Version</code><br/> Returns the version of the currently running application. </li> <li> <code>Initialise Setup</code><br/> Prepares the environment for performing application management operations. </li> <li> <code>Launch App (event_name=None)</code><br/> Launches the default application configured in the session. </li> <li> <code>Start Appium Session (event_name=None)</code><br/> Starts a new Appium session for the current application. </li> <li> <code>Start Other App (package_name, event_name)</code><br/> Launches a different application using the provided package name. </li> </ul> </details> <details> <summary><strong>🔹 FlowControl</strong></summary> <ul> <li> <code>Condition </code><br/> Evaluates multiple conditions and executes corresponding modules if the condition is true. </li> <li> <code>Evaluate (param1, param2)</code><br/> Evaluates a mathematical or logical expression and stores the result in a variable. </li> <li> <code>Read Data (input_element, file_path, index=None)</code><br/> Reads data from a CSV file, API URL, or list and assigns it to a variable. </li> <li> <code>Run Loop (target, *args)</code><br/> Runs a loop either by count or by iterating over variable-value pairs. </li> </ul> </details> <details> <summary><strong>🔹 Verifier</strong></summary> <ul> <li> <code>Assert Equality (output, expression)</code><br/> Compares two values and checks if they are equal. </li> <li> <code>Assert Images Vision (frame, images, element_status, rule)</code><br/> Searches for the specified image templates within the frame using vision-based template matching. </li> <li> <code>Assert Presence (elements, timeout=30, rule='any', event_name=None)</code><br/> Verifies the presence of given elements using Appium or vision-based fallback logic. </li> <li> <code>Assert Texts Vision (frame, texts, element_status, rule)</code><br/> Searches for text in the given frame using OCR and updates element status. </li> <li> <code>Is Element (element, element_state, timeout, event_name)</code><br/> Checks if a given element exists. </li> <li> <code>Validate Element (element, timeout=10, rule='all', event_name=None)</code><br/> Validates if the given element is present on the screen using defined rule and timeout. </li> <li> <code>Validate Screen (elements, timeout=30, rule='any', event_name=None)</code><br/> Validates the presence of a set of elements on a screen using the defined rule. </li> <li> <code>Vision Search (elements, timeout, rule)</code><br/> Performs vision-based search to detect text or image elements in the screen. </li> </ul> </details> ### Setup Development Environment ```bash git clone git@github.com:mozarkai/optics-framework.git cd Optics_Framework pipx install poetry poetry install --with dev ``` ### Running Tests ```bash poetry install --with tests poetry run pytest ``` ### Build Documentation ```bash poetry install --with docs poetry run mkdocs serve ``` ### Packaging the Project ```bash poetry build ``` --- ## 📜 Contributing We welcome contributions! Please follow these steps: 1. Fork the repository. 2. Create a new feature branch. 3. Commit your changes. 4. Open a pull request. Ensure your code follows **PEP8** standards and is formatted with **Black**. --- ## 🎯 Roadmap Here are the key initiatives planned for the upcoming quarter: 1. MCP Servicer: Introduce a dedicated service to handle MCP (Model Context Protocol), improving scalability and modularity across the framework. 2. Omniparser Integration: Seamlessly integrate Omniparser to enable robust and flexible element extraction and location. 3. Playwright Integration: Add support for Playwright to enhance browser automation capabilities, enabling cross-browser testing with modern and powerful tooling. 4. Audio Support: Extend the framework to support audio inputs and outputs, enabling testing and verification of voice-based or sound-related interactions. --- ## 📄 License This project is licensed under the **Apache 2.0 License**. See the [LICENSE](https://github.com/mozarkai/optics-framework?tab=Apache-2.0-1-ov-file) file for details. --- ## 📞 Support For support, please open an issue on GitHub or contact us at [@malto101], [@davidamo9] or [lalit@mozark.ai] . Happy Testing! 🚀
text/markdown
Lalitanand Dandge
lalit@mozark.ai
null
null
Apache-2.0
test automation, framework, mobile automation
[ "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Topic :: Software Development :: Testing" ]
[]
null
null
<4.0,>=3.12
[]
[]
[]
[ "fastapi<0.119.0,>=0.115.12", "fuzzywuzzy<0.19.0,>=0.18.0", "jsonpath-ng<2.0.0,>=1.7.0", "lxml<7.0.0,>=5.3.1", "opencv-python<5.0.0.0,>=4.11.0.86", "pandas<3.0.0,>=2.2.3", "prompt-toolkit<4.0.0,>=3.0.50", "pydantic<3.0.0,>=2.11.2", "pytest<9.0.0,>=8.3.5", "python-json-logger<5.0,>=3.3", "python-levenshtein<0.28.0,>=0.27.1", "pyyaml<7.0.0,>=6.0.2", "requests<3.0.0,>=2.32.3", "rich<15.0.0,>=13.9.4", "scikit-image<0.26.0,>=0.25.2", "sse-starlette<4.0.0,>=2.2.1", "textual<7,>=3", "typing_extensions>=4.13.0", "uvicorn<0.38,>=0.35" ]
[]
[]
[]
[ "Documentation, https://mozarkai.github.io/optics-framework/", "Homepage, https://github.com/mozarkai/optics-framework", "Repository, https://github.com/mozarkai/optics-framework" ]
twine/6.1.0 CPython/3.12.9
2026-02-20T22:07:55.320083
optics_framework-1.8.3.tar.gz
321,693
85/c2/f9d27a279161c7d29e7572a1a64469eb6e5fbb9beab7f43107d39a07ad6c/optics_framework-1.8.3.tar.gz
source
sdist
null
false
f01c687b08853008db8429f1cfd96ec2
4d725b5892567699b59a7c3a8fbeb69b76449d579902f64a9a11d66c014b0950
85c2f9d27a279161c7d29e7572a1a64469eb6e5fbb9beab7f43107d39a07ad6c
null
[ "LICENSE" ]
198
2.4
TruthTorchLM
0.1.19
TruthTorchLM is an open-source library designed to assess truthfulness in language models' outputs. The library integrates state-of-the-art methods, offers comprehensive benchmarking tools across various tasks, and enables seamless integration with popular frameworks like Huggingface and LiteLLM.
<p align="center"> <img align="center" src="https://github.com/Ybakman/TruthTorchLM/blob/main/ttlm_logo.png?raw=true" width="460px" /> </p> <p align="left"> ## TruthTorchLM: A Comprehensive Package for Assessing/Predicting Truthfulness in LLM Outputs (EMNLP - 2025) --- ## Features - **State-of-the-Art Methods**: Offers more than 30 **truth methods** that are designed to assess/predict the truthfulness of LLM generations. These methods range from Google search check to uncertainty estimation and multi-LLM collaboration techniques. - **Integration**: Fully compatible with **Huggingface** and **LiteLLM**, enabling users to integrate truthfulness assessment/prediction into their workflows with **minimal code changes**. - **Evaluation Tools**: Benchmark truth methods using various metrics including AUROC, AUPRC, PRR, and Accuracy. - **Calibration**: Normalize and calibrate truth methods for interpretable and comparable outputs. - **Long-Form Generation**: Adapts truth methods to assess/predict truthfulness in long-form text generations effectively. - **Extendability**: Provides an intuitive interface for implementing new truth methods. --- ## Installation Create a new environment with python >=3.10: ```bash conda create --name truthtorchlm python=3.10 conda activate truthtorchlm ``` Then, install TruthTorchLM using pip: ```bash pip install TruthTorchLM ``` Or, alternatively ```bash git clone https://github.com/Ybakman/TruthTorchLM.git pip -r requirements.txt ``` --- ## Demo Video Available in Youtube https://youtu.be/Bim-6Tv_qU4 ## Quick Start ### Setting Up Credentials ```python import os os.environ["OPENAI_API_KEY"] = 'your_open_ai_key'#to use openai models os.environ['SERPER_API_KEY'] = 'your_serper_api_key'#for long form generation evaluation: https://serper.dev/ ``` ### Setting Up a Model You can define your model and tokenizer using Huggingface or specify an API-based model: ```python from transformers import AutoModelForCausalLM, AutoTokenizer import TruthTorchLM as ttlm import torch # Huggingface model model = AutoModelForCausalLM.from_pretrained( "meta-llama/Meta-Llama-3-8B-Instruct", torch_dtype=torch.bfloat16 ).to('cuda:0') tokenizer = AutoTokenizer.from_pretrained("meta-llama/Meta-Llama-3-8B-Instruct", use_fast=False) # API model api_model = "gpt-4o" ``` ### Generating Text with Truth Values TruthTorchLM generates messages with a truth value, indicating whether the model output is truthful or not. Various methods (called **truth methods**) can be used for this purpose. Each method can have different algorithms and output ranges. Higher truth values generally suggest truthful outputs. This functionality is mostly useful for short-form QA: ```python # Define truth methods lars = ttlm.truth_methods.LARS() confidence = ttlm.truth_methods.Confidence() self_detection = ttlm.truth_methods.SelfDetection(number_of_questions=5) truth_methods = [lars, confidence, self_detection] ``` ```python # Define a chat history chat = [{"role": "system", "content": "You are a helpful assistant. Give short and precise answers."}, {"role": "user", "content": "What is the capital city of France?"}] ``` ```python # Generate text with truth values (Huggingface model) output_hf_model = ttlm.generate_with_truth_value( model=model, tokenizer=tokenizer, messages=chat, truth_methods=truth_methods, max_new_tokens=100, temperature=0.7 ) # Generate text with truth values (API model) output_api_model = ttlm.generate_with_truth_value( model=api_model, messages=chat, truth_methods=truth_methods ) ``` ### Calibrating Truth Methods Truth values for different methods may not be directly comparable. Use the `calibrate_truth_method` function to normalize truth values to a common range for better interpretability. Note that normalized truth value in the output dictionary is meaningless without calibration. ```python model_judge = ttlm.evaluators.ModelJudge('gpt-4o-mini') for truth_method in truth_methods: truth_method.set_normalizer(ttlm.normalizers.IsotonicRegression()) calibration_results = ttlm.calibrate_truth_method( dataset='trivia_qa', model=model, truth_methods=truth_methods, tokenizer=tokenizer, correctness_evaluator=model_judge, size_of_data=1000, max_new_tokens=64 ) ``` ### Evaluating Truth Methods We can evaluate the truth methods with the `evaluate_truth_method` function. We can define different evaluation metrics including AUROC, AUPRC, AUARC, Accuracy, F1, Precision, Recall, PRR: ```python results = ttlm.evaluate_truth_method( dataset='trivia_qa', model=model, truth_methods=truth_methods, eval_metrics=['auroc', 'prr'], tokenizer=tokenizer, size_of_data=1000, correctness_evaluator=model_judge, max_new_tokens=64 ) ``` ### Truthfulness in Long-Form Generation Assigning a single truth value for a long text is neither practical nor useful. TruthTorchLM first decomposes the generated text into short, single-sentence claims and assigns truth values to these claims using claim check methods. The `long_form_generation_with_truth_value` function returns the generated text, decomposed claims, and their truth values. ```python import TruthTorchLM.long_form_generation as LFG from transformers import DebertaForSequenceClassification, DebertaTokenizer #define a decomposition method that breaks the the long text into claims decomposition_method = LFG.decomposition_methods.StructuredDecompositionAPI(model="gpt-4o-mini", decomposition_depth=1) #Utilize API models to decompose text # decomposition_method = LFG.decomposition_methods.StructuredDecompositionLocal(model, tokenizer, decomposition_depth=1) #Utilize HF models to decompose text #entailment model is used by some truth methods and claim check methods model_for_entailment = DebertaForSequenceClassification.from_pretrained('microsoft/deberta-large-mnli').to('cuda:0') tokenizer_for_entailment = DebertaTokenizer.from_pretrained('microsoft/deberta-large-mnli') ``` ```python #define truth methods confidence = ttlm.truth_methods.Confidence() lars = ttlm.truth_methods.LARS() #define the claim check methods that applies truth methods qa_generation = LFG.claim_check_methods.QuestionAnswerGeneration(model="gpt-4o-mini", tokenizer=None, num_questions=2, max_answer_trials=2, truth_methods=[confidence, lars], seed=0, entailment_model=model_for_entailment, entailment_tokenizer=tokenizer_for_entailment) #HF model and tokenizer can also be used, LM is used to generate question #there are some claim check methods that are directly designed for this purpose, not utilizing truth methods ac_entailment = LFG.claim_check_methods.AnswerClaimEntailment( model="gpt-4o-mini", tokenizer=None, num_questions=3, num_answers_per_question=2, entailment_model=model_for_entailment, entailment_tokenizer=tokenizer_for_entailment) #HF model and tokenizer can also be used, LM is used to generate question ``` ```python #define a chat history chat = [{"role": "system", "content": 'You are a helpful assistant. Give brief and precise answers.'}, {"role": "user", "content": f'Who is Ryan Reynolds?'}] #generate a message with a truth value, it's a wrapper fucntion for model.generate in Huggingface output_hf_model = LFG.long_form_generation_with_truth_value(model=model, tokenizer=tokenizer, messages=chat, decomp_method=decomposition_method, claim_check_methods=[qa_generation, ac_entailment], generation_seed=0) #generate a message with a truth value, it's a wrapper fucntion for litellm.completion in litellm output_api_model = LFG.long_form_generation_with_truth_value(model="gpt-4o-mini", messages=chat, decomp_method=decomposition_method, claim_check_methods=[qa_generation, ac_entailment], generation_seed=0, seed=0) ``` ### Evaluation of Truth Methods in Long-Form Generation We can evaluate truth methods on long-form generation by using `evaluate_truth_method_long_form` function. To obtain the correctness of the claims we follow [SAFE paper](https://arxiv.org/pdf/2403.18802). SAFE performs Google search for each claim and assigns labels as supported, unsupported, or irrelevant. We can define different evaluation metrics including AUROC, AUPRC, AUARC, Accuracy, F1, Precision, Recall, PRR. ```python #create safe object that assigns labels to the claims safe = LFG.ClaimEvaluator(rater='gpt-4o-mini', tokenizer = None, max_steps = 5, max_retries = 10, num_searches = 3) #Define metrics sample_level_eval_metrics = ['f1'] #calculate metric over the claims of a question, then average across all the questions dataset_level_eval_metrics = ['auroc', 'prr'] #calculate the metric across all claims ``` ```python results = LFG.evaluate_truth_method_long_form(dataset='longfact_objects', model='gpt-4o-mini', tokenizer=None, sample_level_eval_metrics=sample_level_eval_metrics, dataset_level_eval_metrics=dataset_level_eval_metrics, decomp_method=decomposition_method, claim_check_methods=[qa_generation], claim_evaluator = safe, size_of_data=3, previous_context=[{'role': 'system', 'content': 'You are a helpful assistant. Give precise answers.'}], user_prompt="Question: {question_context}", seed=41, return_method_details = False, return_calim_eval_details=False, wandb_run = None, add_generation_prompt = True, continue_final_message = False) ``` --- ## Available Truth Methods - **LARS**: [Do Not Design, Learn: A Trainable Scoring Function for Uncertainty Estimation in Generative LLMs](https://arxiv.org/pdf/2406.11278). - **Confidence**: [Uncertainty Estimation in Autoregressive Structured Prediction](https://openreview.net/pdf?id=jN5y-zb5Q7m). - **Entropy**:[Uncertainty Estimation in Autoregressive Structured Prediction](https://openreview.net/pdf?id=jN5y-zb5Q7m). - **SelfDetection**: [Knowing What LLMs DO NOT Know: A Simple Yet Effective Self-Detection Method](https://arxiv.org/pdf/2310.17918). - **AttentionScore**: [LLM-Check: Investigating Detection of Hallucinations in Large Language Models](https://openreview.net/pdf?id=LYx4w3CAgy). - **CrossExamination**: [LM vs LM: Detecting Factual Errors via Cross Examination](https://arxiv.org/pdf/2305.13281). - **EccentricityConfidence**: [Generating with Confidence: Uncertainty Quantification for Black-box Large Language Models](https://arxiv.org/pdf/2305.19187). - **EccentricityUncertainty**: [Generating with Confidence: Uncertainty Quantification for Black-box Large Language Models](https://arxiv.org/pdf/2305.19187). - **GoogleSearchCheck**: [FacTool: Factuality Detection in Generative AI -- A Tool Augmented Framework for Multi-Task and Multi-Domain Scenarios](https://arxiv.org/pdf/2307.13528). - **Inside**: [INSIDE: LLMs' Internal States Retain the Power of Hallucination Detection](https://openreview.net/pdf?id=Zj12nzlQbz). - **KernelLanguageEntropy**: [Kernel Language Entropy: Fine-grained Uncertainty Quantification for LLMs from Semantic Similarities](https://arxiv.org/pdf/2405.20003). - **MARS**: [MARS: Meaning-Aware Response Scoring for Uncertainty Estimation in Generative LLMs](https://aclanthology.org/2024.acl-long.419.pdf). - **MatrixDegreeConfidence**: [Generating with Confidence: Uncertainty Quantification for Black-box Large Language Models](https://arxiv.org/pdf/2305.19187). - **MatrixDegreeUncertainty**: [Generating with Confidence: Uncertainty Quantification for Black-box Large Language Models](https://arxiv.org/pdf/2305.19187). - **MiniCheck**: [MiniCheck: Efficient Fact-Checking of LLMs on Grounding Documents](https://arxiv.org/pdf/2404.10774) - **MultiLLMCollab**: [Don’t Hallucinate, Abstain: Identifying LLM Knowledge Gaps via Multi-LLM Collaboration](https://arxiv.org/pdf/2402.00367). - **NumSemanticSetUncertainty**: [Semantic Uncertainty: Linguistic Invariances for Uncertainty Estimation in Natural Language Generation](https://arxiv.org/pdf/2302.09664). - **PTrue**: [Language Models (Mostly) Know What They Know](https://arxiv.org/pdf/2207.05221). - **Saplma**: [The Internal State of an LLM Knows When It’s Lying](https://aclanthology.org/2023.findings-emnlp.68.pdf). - **SemanticEntropy**: [Semantic Uncertainty: Linguistic Invariances for Uncertainty Estimation in Natural Language Generation](https://arxiv.org/pdf/2302.09664). - **sentSAR**: [Shifting Attention to Relevance: Towards the Predictive Uncertainty Quantification of Free-Form Large Language Models](https://aclanthology.org/2024.acl-long.276.pdf). - **SumEigenUncertainty**: [Generating with Confidence: Uncertainty Quantification for Black-box Large Language Models](https://arxiv.org/pdf/2305.19187). - **tokenSAR**: [Shifting Attention to Relevance: Towards the Predictive Uncertainty Quantification of Free-Form Large Language Models](https://aclanthology.org/2024.acl-long.276.pdf). - **VerbalizedConfidence**: [Just Ask for Calibration: Strategies for Eliciting Calibrated Confidence Scores from Language Models Fine-Tuned with Human Feedback](https://openreview.net/pdf?id=g3faCfrwm7). - **DirectionalEntailmentGraph**: [LLM Uncertainty Quantification through Directional Entailment Graph and Claim Level Response Augmentation](https://arxiv.org/pdf/2407.00994) --- ## Contributors - **Yavuz Faruk Bakman** (ybakman@usc.edu) - **Duygu Nur Yaldiz** (yaldiz@usc.edu) - **Sungmin Kang** (kangsung@usc.edu) - **Alperen Ozis** (alperenozis@gmail.com) - **Hayrettin Eren Yildiz** (hayereyil@gmail.com) - **Mitash Shah** (mitashsh@usc.edu) --- ## Citation If you use TruthTorchLM in your research, please cite: ```bibtex @inproceedings{yaldiz-etal-2025-truthtorchlm, title = "{T}ruth{T}orch{LM}: A Comprehensive Library for Predicting Truthfulness in {LLM} Outputs", author = {Yaldiz, Duygu Nur and Bakman, Yavuz Faruk and Kang, Sungmin and {\"O}zi{\c{s}}, Alperen and Yildiz, Hayrettin Eren and Shah, Mitash Ashish and Huang, Zhiqi and Kumar, Anoop and Samuel, Alfy and Liu, Daben and Karimireddy, Sai Praneeth and Avestimehr, Salman}, editor = {Habernal, Ivan and Schulam, Peter and Tiedemann, J{\"o}rg}, booktitle = "Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing: System Demonstrations", month = nov, year = "2025", address = "Suzhou, China", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/2025.emnlp-demos.54/", pages = "717--728", ISBN = "979-8-89176-334-0", } ``` --- ## License TruthTorchLM is released under the [MIT License](LICENSE). For inquiries or support, feel free to contact the maintainers.
text/markdown
Yavuz Faruk Bakman
ybakman@usc.edu
null
null
null
null
[]
[]
null
null
>=3.10
[]
[]
[]
[ "aiohttp", "evaluate", "instructor", "litellm", "nest_asyncio", "numpy", "outlines", "pandas", "pydantic", "PyYAML", "Requests", "scikit_learn", "scipy", "sentence_transformers", "termcolor", "torch", "tqdm", "transformers", "absl-py", "nltk", "rouge_score", "wandb", "sentencepiece", "accelerate>=0.26.0" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-20T22:06:04.364371
truthtorchlm-0.1.19.tar.gz
93,093
16/2f/4d8faa846f54c6b80149096bb276893669099ccccfb86047927de0b83211/truthtorchlm-0.1.19.tar.gz
source
sdist
null
false
daa6a673c764adbb69ba3259c344b88e
f708056c5cc2d38cfe268e20785dfd4d7f3e4d6e54e4e7d13461bafdb5590677
162f4d8faa846f54c6b80149096bb276893669099ccccfb86047927de0b83211
null
[ "LICENSE" ]
0
2.4
ziphq-mcp
0.1.4
MCP server for Zip API integration
# Zip MCP Server An MCP (Model Context Protocol) server for integrating with the Zip API, built with [FastMCP](https://github.com/jlowin/fastmcp). ## Prerequisite Make sure you install `uvx` globally from [here](https://docs.astral.sh/uv/getting-started/installation/) before proceeding. ## Usage ### With MCP client Add to your MCP configuration (e.g., `~/.claude/claude_desktop_config.json`): ```json { "mcpServers": { "zip": { "command": "uvx", "args": ["ziphq-mcp"], "env": { "ZIP_API_KEY": "your-api-key-here" } } } }
text/markdown
null
Hai Nguyen <hai@ziphq.com>
null
null
null
null
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12" ]
[]
null
null
>=3.10
[]
[]
[]
[ "fastmcp>=2.0.0", "httpx>=0.28.0", "pytest-asyncio>=0.23.0; extra == \"dev\"", "pytest>=8.0.0; extra == \"dev\"", "ruff>=0.3.0; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://github.com/Greenbax/evergreen", "Repository, https://github.com/Greenbax/evergreen" ]
uv/0.9.6
2026-02-20T22:04:47.445888
ziphq_mcp-0.1.4.tar.gz
87,857
dc/65/d8ed7d9a199a0bb858fea7f9f050df95df65fc8b42b7a700dbf1dc90eaca/ziphq_mcp-0.1.4.tar.gz
source
sdist
null
false
d58c835746a98659bc2b9ee893150c57
8b1848d751cd79b070b96c2f6d6fe28218b2ca8c56d7764489cbde3f59d88f20
dc65d8ed7d9a199a0bb858fea7f9f050df95df65fc8b42b7a700dbf1dc90eaca
MIT
[ "LICENSE" ]
199
2.4
mcp-codebase-index
0.4.5
Structural codebase indexer with MCP server for AI-assisted development
<!-- mcp-name: io.github.MikeRecognex/mcp-codebase-index --> # mcp-codebase-index [![PyPI version](https://img.shields.io/pypi/v/mcp-codebase-index)](https://pypi.org/project/mcp-codebase-index/) [![CI](https://github.com/MikeRecognex/mcp-codebase-index/actions/workflows/ci.yml/badge.svg)](https://github.com/MikeRecognex/mcp-codebase-index/actions/workflows/ci.yml) [![Python 3.11+](https://img.shields.io/badge/python-3.11+-blue.svg)](https://www.python.org/downloads/) [![License: AGPL-3.0](https://img.shields.io/badge/License-AGPL--3.0-blue.svg)](https://www.gnu.org/licenses/agpl-3.0) [![MCP](https://img.shields.io/badge/MCP-compatible-purple.svg)](https://modelcontextprotocol.io) [![Zero Dependencies](https://img.shields.io/badge/dependencies-zero-brightgreen.svg)]() A structural codebase indexer with an [MCP](https://modelcontextprotocol.io) server for AI-assisted development. Zero runtime dependencies — uses Python's `ast` module for Python analysis and regex-based parsing for TypeScript/JS, Go, and Rust. Requires Python 3.11+. ## What It Does Indexes codebases by parsing source files into structural metadata -- functions, classes, imports, dependency graphs, and cross-file call chains -- then exposes 18 query tools via the Model Context Protocol, enabling Claude Code and other MCP clients to navigate codebases efficiently without reading entire files. **Automatic incremental re-indexing:** In git repositories, the index stays up to date automatically. Before every query, the server checks `git diff` and `git status` (~1-2ms). If files changed, only those files are re-parsed and the dependency graph is rebuilt. No need to manually call `reindex` after edits, branch switches, or pulls. ## Language Support | Language | Method | Extracts | |----------|--------|----------| | Python (`.py`) | AST parsing | Functions, classes, methods, imports, dependency graph | | TypeScript/JS (`.ts`, `.tsx`, `.js`, `.jsx`) | Regex-based | Functions, arrow functions, classes, interfaces, type aliases, imports | | Go (`.go`) | Regex-based | Functions, methods (receiver-based), structs, interfaces, type aliases, imports, doc comments | | Rust (`.rs`) | Regex-based | Functions (`pub`/`async`/`const`/`unsafe`), structs, enums, traits, impl blocks, use statements, attributes, doc comments, macro_rules | | Markdown/Text (`.md`, `.txt`, `.rst`) | Heading detection | Sections (# headings, underlines, numbered, ALL-CAPS) | | Other | Generic | Line counts only | ## Installation ```bash pip install "mcp-codebase-index[mcp]" ``` The `[mcp]` extra includes the MCP server dependency. Omit it if you only need the programmatic API. For development (from a local clone): ```bash pip install -e ".[dev,mcp]" ``` ## MCP Server ### Running ```bash # As a console script PROJECT_ROOT=/path/to/project mcp-codebase-index # As a Python module PROJECT_ROOT=/path/to/project python -m mcp_codebase_index.server ``` `PROJECT_ROOT` specifies which directory to index. Defaults to the current working directory. ### Configuring with OpenClaw Install the package on the machine where OpenClaw is running: ```bash # Local install pip install "mcp-codebase-index[mcp]" # Or inside a Docker container / remote VPS docker exec -it openclaw bash pip install "mcp-codebase-index[mcp]" ``` Add the MCP server to your OpenClaw agent config (`openclaw.json`): ```json { "agents": { "list": [{ "id": "main", "mcp": { "servers": [ { "name": "codebase-index", "command": "mcp-codebase-index", "env": { "PROJECT_ROOT": "/path/to/project" } } ] } }] } } ``` Restart OpenClaw and verify the connection: ```bash openclaw mcp list ``` All 18 tools will be available to your agent. **Performance note:** The server automatically detects file changes via `git diff` before every query (~1-2ms) and incrementally re-indexes only what changed. However, OpenClaw's default MCP integration via mcporter spawns a fresh server process per tool call, which discards the in-memory index and forces a full rebuild each time (~1-2s for small projects, longer for large ones). This is a mcporter process lifecycle limitation, not a server limitation. For persistent connections, use the [openclaw-mcp-adapter](https://github.com/androidStern-personal/openclaw-mcp-adapter) plugin, which connects once at startup and keeps the server running: ```bash pip install openclaw-mcp-adapter ``` ### Configuring with Claude Code Add to your project's `.mcp.json`: ```json { "mcpServers": { "codebase-index": { "command": "mcp-codebase-index", "env": { "PROJECT_ROOT": "/path/to/project" } } } } ``` Or using the Python module directly (useful if installed in a virtualenv): ```json { "mcpServers": { "codebase-index": { "command": "/path/to/.venv/bin/python3", "args": ["-m", "mcp_codebase_index.server"], "env": { "PROJECT_ROOT": "/path/to/project" } } } } ``` ### Important: Make the AI Actually Use Indexed Tools By default, AI assistants will ignore the indexed tools and fall back to reading entire files with Glob/Grep/Read. Soft language like "prefer" gets rationalized away. Add this to your project's `CLAUDE.md` (or equivalent instructions file) with **mandatory** language: ``` ## Codebase Navigation — MANDATORY You MUST use codebase-index MCP tools FIRST when exploring or navigating the codebase. This is not optional. - ALWAYS start with: get_project_summary, find_symbol, get_function_source, get_class_source, get_structure_summary, get_dependencies, get_dependents, get_change_impact, get_call_chain, search_codebase - Only fall back to Read/Glob/Grep when codebase-index tools genuinely don't have what you need (e.g. reading non-code files, config, frontmatter) - If you catch yourself reaching for Glob/Grep/Read to find or understand code, STOP and use codebase-index instead ``` The word "prefer" is too weak — models treat it as a suggestion and default to familiar tools. Mandatory language with explicit fallback criteria is what actually changes behavior. ### Available Tools (18) | Tool | Description | |------|-------------| | `get_project_summary` | File count, packages, top classes/functions | | `list_files` | List indexed files with optional glob filter | | `get_structure_summary` | Structure of a file or the whole project | | `get_functions` | List functions with name, lines, params | | `get_classes` | List classes with name, lines, methods, bases | | `get_imports` | List imports with module, names, line | | `get_function_source` | Full source of a function/method | | `get_class_source` | Full source of a class | | `find_symbol` | Find where a symbol is defined (file, line, type) | | `get_dependencies` | What a symbol calls/uses | | `get_dependents` | What calls/uses a symbol | | `get_change_impact` | Direct + transitive dependents | | `get_call_chain` | Shortest dependency path (BFS) | | `get_file_dependencies` | Files imported by a given file | | `get_file_dependents` | Files that import from a given file | | `search_codebase` | Regex search across all files (max 100 results) | | `reindex` | Force full re-index (rarely needed — incremental updates happen automatically in git repos) | | `get_usage_stats` | Session efficiency stats: tool calls, characters returned vs total source, estimated token savings | ## Benchmarks Tested across four real-world projects on an M-series MacBook Pro, from a small project to CPython itself (1.1 million lines): ### Index Build Performance | Project | Files | Lines | Functions | Classes | Index Time | Peak Memory | |---------|------:|------:|----------:|--------:|-----------:|------------:| | RMLPlus | 36 | 7,762 | 237 | 55 | 0.9s | 2.4 MB | | FastAPI | 2,556 | 332,160 | 4,139 | 617 | 5.7s | 55 MB | | Django | 3,714 | 707,493 | 29,995 | 7,371 | 36.2s | 126 MB | | **CPython** | **2,464** | **1,115,334** | **59,620** | **9,037** | **55.9s** | **197 MB** | ### Query Response Size vs Total Source Querying CPython — 41 million characters of source code: | Query | Response | Total Source | Reduction | |-------|-------:|------------:|----------:| | `find_symbol("TestCase")` | 67 chars | 41,077,561 chars | **99.9998%** | | `get_dependencies("compile")` | 115 chars | 41,077,561 chars | **99.9997%** | | `get_change_impact("TestCase")` | 16,812 chars | 41,077,561 chars | **99.96%** | | `get_function_source("compile")` | 4,531 chars | 41,077,561 chars | **99.99%** | | `get_function_source("run_unittest")` | 439 chars | 41,077,561 chars | **99.999%** | `find_symbol` returns 54-67 characters regardless of whether the project is 7K lines or 1.1M lines. Response size scales with the answer, not the codebase. `get_change_impact("TestCase")` on CPython found **154 direct dependents and 492 transitive dependents** in 0.45ms — the kind of query that's impossible without a dependency graph. Use `max_direct` and `max_transitive` to cap output to your token budget. ### Query Response Time All targeted queries return in sub-millisecond time, even on CPython's 1.1M lines: | Query | RMLPlus | FastAPI | Django | CPython | |-------|--------:|--------:|-------:|--------:| | `find_symbol` | 0.01ms | 0.01ms | 0.03ms | 0.08ms | | `get_dependencies` | 0.00ms | 0.00ms | 0.00ms | 0.01ms | | `get_change_impact` | 0.02ms | 0.00ms | 2.81ms | 0.45ms | | `get_function_source` | 0.01ms | 0.02ms | 0.03ms | 0.10ms | Run the benchmarks yourself: `python benchmarks/benchmark.py` ## How Is This Different from LSP? LSP answers "where is this function?" — mcp-codebase-index answers "what happens if I change it?" LSP is point queries: one symbol, one file, one position. It can tell you where `LLMClient` is defined and who references it. But ask "what breaks transitively if I refactor `LLMClient`?" and LSP has nothing. This tool returns 11 direct dependents and 31 transitive impacts in a single call — 204 characters. To get the same answer from LSP, the AI would need to chain dozens of find-reference calls recursively, reading files at every step, burning thousands of tokens to reconstruct what the dependency graph already knows. LSP also requires you to install a separate language server for every language in your project — pyright for Python, vtsls for TypeScript, gopls for Go. Each one is a heavyweight binary with its own dependencies and configuration. mcp-codebase-index is zero dependencies, handles Python + TypeScript/JS + Go + Rust + Markdown out of the box, and every response has built-in token budget controls (`max_results`, `max_lines`). LSP was built for IDEs. This was built for AI. ## Programmatic Usage ```python from mcp_codebase_index.project_indexer import ProjectIndexer from mcp_codebase_index.query_api import create_project_query_functions indexer = ProjectIndexer("/path/to/project", include_patterns=["**/*.py"]) index = indexer.index() query_funcs = create_project_query_functions(index) # Use query functions print(query_funcs["get_project_summary"]()) print(query_funcs["find_symbol"]("MyClass")) print(query_funcs["get_change_impact"]("some_function")) ``` ## Development ```bash pip install -e ".[dev,mcp]" pytest tests/ -v ruff check src/ tests/ ``` ## References The structural indexer was originally developed as part of the [RMLPlus](https://github.com/MikeRecognex/RMLPlus) project, an implementation of the [Recursive Language Models](https://arxiv.org/abs/2512.24601) framework. ## License This project is dual-licensed: - **AGPL-3.0** for open-source use — see [LICENSE](LICENSE) - **Commercial License** for proprietary use — see [COMMERCIAL-LICENSE.md](COMMERCIAL-LICENSE.md) If you're using mcp-codebase-index as a standalone MCP server for development, the AGPL-3.0 license applies at no cost. If you're embedding it in a proprietary product or offering it as part of a hosted service, you'll need a commercial license. See [COMMERCIAL-LICENSE.md](COMMERCIAL-LICENSE.md) for details.
text/markdown
Michael Doyle
null
null
null
GNU AFFERO GENERAL PUBLIC LICENSE Version 3, 19 November 2007 Copyright (C) 2007 Free Software Foundation, Inc. <https://fsf.org/> Everyone is permitted to copy and distribute verbatim copies of this license document, but changing it is not allowed. Preamble The GNU Affero General Public License is a free, copyleft license for software and other kinds of works, specifically designed to ensure cooperation with the community in the case of network server software. The licenses for most software and other practical works are designed to take away your freedom to share and change the works. By contrast, our General Public Licenses are intended to guarantee your freedom to share and change all versions of a program--to make sure it remains free software for all its users. When we speak of free software, we are referring to freedom, not price. Our General Public Licenses are designed to make sure that you have the freedom to distribute copies of free software (and charge for them if you wish), that you receive source code or can get it if you want it, that you can change the software or use pieces of it in new free programs, and that you know you can do these things. Developers that use our General Public Licenses protect your rights with two steps: (1) assert copyright on the software, and (2) offer you this License which gives you legal permission to copy, distribute and/or modify the software. A secondary benefit of defending all users' freedom is that improvements made in alternate versions of the program, if they receive widespread use, become available for other developers to incorporate. Many developers of free software are heartened and encouraged by the resulting cooperation. However, in the case of software used on network servers, this result may fail to come about. The GNU General Public License permits making a modified version and letting the public access it on a server without ever releasing its source code to the public. The GNU Affero General Public License is designed specifically to ensure that, in such cases, the modified source code becomes available to the community. It requires the operator of a network server to provide the source code of the modified version running there to the users of that server. Therefore, public use of a modified version, on a publicly accessible server, gives the public access to the source code of the modified version. An older license, called the Affero General Public License and published by Affero, was designed to accomplish similar goals. This is a different license, not a version of the Affero GPL, but Affero has released a new version of the Affero GPL which permits relicensing under this license. The precise terms and conditions for copying, distribution and modification follow. TERMS AND CONDITIONS 0. Definitions. "This License" refers to version 3 of the GNU Affero General Public License. "Copyright" also means copyright-like laws that apply to other kinds of works, such as semiconductor masks. "The Program" refers to any copyrightable work licensed under this License. Each licensee is addressed as "you". "Licensees" and "recipients" may be individuals or organizations. To "modify" a work means to copy from or adapt all or part of the work in a fashion requiring copyright permission, other than the making of an exact copy. The resulting work is called a "modified version" of the earlier work or a work "based on" the earlier work. A "covered work" means either the unmodified Program or a work based on the Program. To "propagate" a work means to do anything with it that, without permission, would make you directly or secondarily liable for infringement under applicable copyright law, except executing it on a computer or modifying a private copy. Propagation includes copying, distribution (with or without modification), making available to the public, and in some countries other activities as well. To "convey" a work means any kind of propagation that enables other parties to make or receive copies. Mere interaction with a user through a computer network, with no transfer of a copy, is not conveying. An interactive user interface displays "Appropriate Legal Notices" to the extent that it includes a convenient and prominently visible feature that (1) displays an appropriate copyright notice, and (2) tells the user that there is no warranty for the work (except to the extent that warranties are provided), that licensees may convey the work under this License, and how to view a copy of this License. If the interface presents a list of user commands or options, such as a menu, a prominent item in the list meets this criterion. 1. Source Code. The "source code" for a work means the preferred form of the work for making modifications to it. "Object code" means any non-source form of a work. A "Standard Interface" means an interface that either is an official standard defined by a recognized standards body, or, in the case of interfaces specified for a particular programming language, one that is widely used among developers working in that language. The "System Libraries" of an executable work include anything, other than the work as a whole, that (a) is included in the normal form of packaging a Major Component, but which is not part of that Major Component, and (b) serves only to enable use of the work with that Major Component, or to implement a Standard Interface for which an implementation is available to the public in source code form. A "Major Component", in this context, means a major essential component (kernel, window system, and so on) of the specific operating system (if any) on which the executable work runs, or a compiler used to produce the work, or an object code interpreter used to run it. The "Corresponding Source" for a work in object code form means all the source code needed to generate, install, and (for an executable work) run the object code and to modify the work, including scripts to control those activities. However, it does not include the work's System Libraries, or general-purpose tools or generally available free programs which are used unmodified in performing those activities but which are not part of the work. For example, Corresponding Source includes interface definition files associated with source files for the work, and the source code for shared libraries and dynamically linked subprograms that the work is specifically designed to require, such as by intimate data communication or control flow between those subprograms and other parts of the work. The Corresponding Source need not include anything that users can regenerate automatically from other parts of the Corresponding Source. The Corresponding Source for a work in source code form is that same work. 2. Basic Permissions. All rights granted under this License are granted for the term of copyright on the Program, and are irrevocable provided the stated conditions are met. This License explicitly affirms your unlimited permission to run the unmodified Program. The output from running a covered work is covered by this License only if the output, given its content, constitutes a covered work. This License acknowledges your rights of fair use or other equivalent, as provided by copyright law. You may make, run and propagate covered works that you do not convey, without conditions so long as your license otherwise remains in force. You may convey covered works to others for the sole purpose of having them make modifications exclusively for you, or provide you with facilities for running those works, provided that you comply with the terms of this License in conveying all material for which you do not control copyright. Those thus making or running the covered works for you must do so exclusively on your behalf, under your direction and control, on terms that prohibit them from making any copies of your copyrighted material outside their relationship with you. Conveying under any other circumstances is permitted solely under the conditions stated below. Sublicensing is not allowed; section 10 makes it unnecessary. 3. Protecting Users' Legal Rights From Anti-Circumvention Law. No covered work shall be deemed part of an effective technological measure under any applicable law fulfilling obligations under article 11 of the WIPO copyright treaty adopted on 20 December 1996, or similar laws prohibiting or restricting circumvention of such measures. When you convey a covered work, you waive any legal power to forbid circumvention of technological measures to the extent such circumvention is effected by exercising rights under this License with respect to the covered work, and you disclaim any intention to limit operation or modification of the work as a means of enforcing, against the work's users, your or third parties' legal rights to forbid circumvention of technological measures. 4. Conveying Verbatim Copies. You may convey verbatim copies of the Program's source code as you receive it, in any medium, provided that you conspicuously and appropriately publish on each copy an appropriate copyright notice; keep intact all notices stating that this License and any non-permissive terms added in accord with section 7 apply to the code; keep intact all notices of the absence of any warranty; and give all recipients a copy of this License along with the Program. You may charge any price or no price for each copy that you convey, and you may offer support or warranty protection for a fee. 5. Conveying Modified Source Versions. You may convey a work based on the Program, or the modifications to produce it from the Program, in the form of source code under the terms of section 4, provided that you also meet all of these conditions: a) The work must carry prominent notices stating that you modified it, and giving a relevant date. b) The work must carry prominent notices stating that it is released under this License and any conditions added under section 7. This requirement modifies the requirement in section 4 to "keep intact all notices". c) You must license the entire work, as a whole, under this License to anyone who comes into possession of a copy. This License will therefore apply, along with any applicable section 7 additional terms, to the whole of the work, and all its parts, regardless of how they are packaged. This License gives no permission to license the work in any other way, but it does not invalidate such permission if you have separately received it. d) If the work has interactive user interfaces, each must display Appropriate Legal Notices; however, if the Program has interactive interfaces that do not display Appropriate Legal Notices, your work need not make them do so. A compilation of a covered work with other separate and independent works, which are not by their nature extensions of the covered work, and which are not combined with it such as to form a larger program, in or on a volume of a storage or distribution medium, is called an "aggregate" if the compilation and its resulting copyright are not used to limit the access or legal rights of the compilation's users beyond what the individual works permit. Inclusion of a covered work in an aggregate does not cause this License to apply to the other parts of the aggregate. 6. Conveying Non-Source Forms. You may convey a covered work in object code form under the terms of sections 4 and 5, provided that you also convey the machine-readable Corresponding Source under the terms of this License, in one of these ways: a) Convey the object code in, or embodied in, a physical product (including a physical distribution medium), accompanied by the Corresponding Source fixed on a durable physical medium customarily used for software interchange. b) Convey the object code in, or embodied in, a physical product (including a physical distribution medium), accompanied by a written offer, valid for at least three years and valid for as long as you offer spare parts or customer support for that product model, to give anyone who possesses the object code either (1) a copy of the Corresponding Source for all the software in the product that is covered by this License, on a durable physical medium customarily used for software interchange, for a price no more than your reasonable cost of physically performing this conveying of source, or (2) access to copy the Corresponding Source from a network server at no charge. c) Convey individual copies of the object code with a copy of the written offer to provide the Corresponding Source. This alternative is allowed only occasionally and noncommercially, and only if you received the object code with such an offer, in accord with subsection 6b. d) Convey the object code by offering access from a designated place (gratis or for a charge), and offer equivalent access to the Corresponding Source in the same way through the same place at no further charge. You need not require recipients to copy the Corresponding Source along with the object code. If the place to copy the object code is a network server, the Corresponding Source may be on a different server (operated by you or a third party) that supports equivalent copying facilities, provided you maintain clear directions next to the object code saying where to find the Corresponding Source. Regardless of what server hosts the Corresponding Source, you remain obligated to ensure that it is available for as long as needed to satisfy these requirements. e) Convey the object code using peer-to-peer transmission, provided you inform other peers where the object code and Corresponding Source of the work are being offered to the general public at no charge under subsection 6d. A separable portion of the object code, whose source code is excluded from the Corresponding Source as a System Library, need not be included in conveying the object code work. A "User Product" is either (1) a "consumer product", which means any tangible personal property which is normally used for personal, family, or household purposes, or (2) anything designed or sold for incorporation into a dwelling. In determining whether a product is a consumer product, doubtful cases shall be resolved in favor of coverage. For a particular product received by a particular user, "normally used" refers to a typical or common use of that class of product, regardless of the status of the particular user or of the way in which the particular user actually uses, or expects or is expected to use, the product. A product is a consumer product regardless of whether the product has substantial commercial, industrial or non-consumer uses, unless such uses represent the only significant mode of use of the product. "Installation Information" for a User Product means any methods, procedures, authorization keys, or other information required to install and execute modified versions of a covered work in that User Product from a modified version of its Corresponding Source. The information must suffice to ensure that the continued functioning of the modified object code is in no case prevented or interfered with solely because modification has been made. If you convey an object code work under this section in, or with, or specifically for use in, a User Product, and the conveying occurs as part of a transaction in which the right of possession and use of the User Product is transferred to the recipient in perpetuity or for a fixed term (regardless of how the transaction is characterized), the Corresponding Source conveyed under this section must be accompanied by the Installation Information. But this requirement does not apply if neither you nor any third party retains the ability to install modified object code on the User Product (for example, the work has been installed in ROM). The requirement to provide Installation Information does not include a requirement to continue to provide support service, warranty, or updates for a work that has been modified or installed by the recipient, or for the User Product in which it has been modified or installed. Access to a network may be denied when the modification itself materially and adversely affects the operation of the network or violates the rules and protocols for communication across the network. Corresponding Source conveyed, and Installation Information provided, in accord with this section must be in a format that is publicly documented (and with an implementation available to the public in source code form), and must require no special password or key for unpacking, reading or copying. 7. Additional Terms. "Additional permissions" are terms that supplement the terms of this License by making exceptions from one or more of its conditions. Additional permissions that are applicable to the entire Program shall be treated as though they were included in this License, to the extent that they are valid under applicable law. If additional permissions apply only to part of the Program, that part may be used separately under those permissions, but the entire Program remains governed by this License without regard to the additional permissions. When you convey a copy of a covered work, you may at your option remove any additional permissions from that copy, or from any part of it. (Additional permissions may be written to require their own removal in certain cases when you modify the work.) You may place additional permissions on material, added by you to a covered work, for which you have or can give appropriate copyright permission. Notwithstanding any other provision of this License, for material you add to a covered work, you may (if authorized by the copyright holders of that material) supplement the terms of this License with terms: a) Disclaiming warranty or limiting liability differently from the terms of sections 15 and 16 of this License; or b) Requiring preservation of specified reasonable legal notices or author attributions in that material or in the Appropriate Legal Notices displayed by works containing it; or c) Prohibiting misrepresentation of the origin of that material, or requiring that modified versions of such material be marked in reasonable ways as different from the original version; or d) Limiting the use for publicity purposes of names of licensors or authors of the material; or e) Declining to grant rights under trademark law for use of some trade names, trademarks, or service marks; or f) Requiring indemnification of licensors and authors of that material by anyone who conveys the material (or modified versions of it) with contractual assumptions of liability to the recipient, for any liability that these contractual assumptions directly impose on those licensors and authors. All other non-permissive additional terms are considered "further restrictions" within the meaning of section 10. If the Program as you received it, or any part of it, contains a notice stating that it is governed by this License along with a term that is a further restriction, you may remove that term. If a license document contains a further restriction but permits relicensing or conveying under this License, you may add to a covered work material governed by the terms of that license document, provided that the further restriction does not survive such relicensing or conveying. If you add terms to a covered work in accord with this section, you must place, in the relevant source files, a statement of the additional terms that apply to those files, or a notice indicating where to find the applicable terms. Additional terms, permissive or non-permissive, may be stated in the form of a separately written license, or stated as exceptions; the above requirements apply either way. 8. Termination. You may not propagate or modify a covered work except as expressly provided under this License. Any attempt otherwise to propagate or modify it is void, and will automatically terminate your rights under this License (including any patent licenses granted under the third paragraph of section 11). However, if you cease all violation of this License, then your license from a particular copyright holder is reinstated (a) provisionally, unless and until the copyright holder explicitly and finally terminates your license, and (b) permanently, if the copyright holder fails to notify you of the violation by some reasonable means prior to 60 days after the cessation. Moreover, your license from a particular copyright holder is reinstated permanently if the copyright holder notifies you of the violation by some reasonable means, this is the first time you have received notice of violation of this License (for any work) from that copyright holder, and you cure the violation prior to 30 days after your receipt of the notice. Termination of your rights under this section does not terminate the licenses of parties who have received copies or rights from you under this License. If your rights have been terminated and not permanently reinstated, you do not qualify to receive new licenses for the same material under section 10. 9. Acceptance Not Required for Having Copies. You are not required to accept this License in order to receive or run a copy of the Program. Ancillary propagation of a covered work occurring solely as a consequence of using peer-to-peer transmission to receive a copy likewise does not require acceptance. However, nothing other than this License grants you permission to propagate or modify any covered work. These actions infringe copyright if you do not accept this License. Therefore, by modifying or propagating a covered work, you indicate your acceptance of this License to do so. 10. Automatic Licensing of Downstream Recipients. Each time you convey a covered work, the recipient automatically receives a license from the original licensors, to run, modify and propagate that work, subject to this License. You are not responsible for enforcing compliance by third parties with this License. An "entity transaction" is a transaction transferring control of an organization, or substantially all assets of one, or subdividing an organization, or merging organizations. If propagation of a covered work results from an entity transaction, each party to that transaction who receives a copy of the work also receives whatever licenses to the work the party's predecessor in interest had or could give under the previous paragraph, plus a right to possession of the Corresponding Source of the work from the predecessor in interest, if the predecessor has it or can get it with reasonable efforts. You may not impose any further restrictions on the exercise of the rights granted or affirmed under this License. For example, you may not impose a license fee, royalty, or other charge for exercise of rights granted under this License, and you may not initiate litigation (including a cross-claim or counterclaim in a lawsuit) alleging that any patent claim is infringed by making, using, selling, offering for sale, or importing the Program or any portion of it. 11. Patents. A "contributor" is a copyright holder who authorizes use under this License of the Program or a work on which the Program is based. The work thus licensed is called the contributor's "contributor version". A contributor's "essential patent claims" are all patent claims owned or controlled by the contributor, whether already acquired or hereafter acquired, that would be infringed by some manner, permitted by this License, of making, using, or selling its contributor version, but do not include claims that would be infringed only as a consequence of further modification of the contributor version. For purposes of this definition, "control" includes the right to grant patent sublicenses in a manner consistent with the requirements of this License. Each contributor grants you a non-exclusive, worldwide, royalty-free patent license under the contributor's essential patent claims, to make, use, sell, offer for sale, import and otherwise run, modify and propagate the contents of its contributor version. In the following three paragraphs, a "patent license" is any express agreement or commitment, however denominated, not to enforce a patent (such as an express permission to practice a patent or covenant not to sue for patent infringement). To "grant" such a patent license to a party means to make such an agreement or commitment not to enforce a patent against the party. If you convey a covered work, knowingly relying on a patent license, and the Corresponding Source of the work is not available for anyone to copy, free of charge and under the terms of this License, through a publicly available network server or other readily accessible means, then you must either (1) cause the Corresponding Source to be so available, or (2) arrange to deprive yourself of the benefit of the patent license for this particular work, or (3) arrange, in a manner consistent with the requirements of this License, to extend the patent license to downstream recipients. "Knowingly relying" means you have actual knowledge that, but for the patent license, your conveying the covered work in a country, or your recipient's use of the covered work in a country, would infringe one or more identifiable patents in that country that you have reason to believe are valid. If, pursuant to or in connection with a single transaction or arrangement, you convey, or propagate by procuring conveyance of, a covered work, and grant a patent license to some of the parties receiving the covered work authorizing them to use, propagate, modify or convey a specific copy of the covered work, then the patent license you grant is automatically extended to all recipients of the covered work and works based on it. A patent license is "discriminatory" if it does not include within the scope of its coverage, prohibits the exercise of, or is conditioned on the non-exercise of one or more of the rights that are specifically granted under this License. You may not convey a covered work if you are a party to an arrangement with a third party that is in the business of distributing software, under which you make payment to the third party based on the extent of your activity of conveying the work, and under which the third party grants, to any of the parties who would receive the covered work from you, a discriminatory patent license (a) in connection with copies of the covered work conveyed by you (or copies made from those copies), or (b) primarily for and in connection with specific products or compilations that contain the covered work, unless you entered into that arrangement, or that patent license was granted, prior to 28 March 2007. Nothing in this License shall be construed as excluding or limiting any implied license or other defenses to infringement that may otherwise be available to you under applicable patent law. 12. No Surrender of Others' Freedom. If conditions are imposed on you (whether by court order, agreement or otherwise) that contradict the conditions of this License, they do not excuse you from the conditions of this License. If you cannot convey a covered work so as to satisfy simultaneously your obligations under this License and any other pertinent obligations, then as a consequence you may not convey it at all. For example, if you agree to terms that obligate you to collect a royalty for further conveying from those to whom you convey the Program, the only way you could satisfy both those terms and this License would be to refrain entirely from conveying the Program. 13. Remote Network Interaction; Use with the GNU General Public License. Notwithstanding any other provision of this License, if you modify the Program, your modified version must prominently offer all users interacting with it remotely through a computer network (if your version supports such interaction) an opportunity to receive the Corresponding Source of your version by providing access to the Corresponding Source from a network server at no charge, through some standard or customary means of facilitating copying of software. This Corresponding Source shall include the Corresponding Source for any work covered by version 3 of the GNU General Public License that is incorporated pursuant to the following paragraph. Notwithstanding any other provision of this License, you have permission to link or combine any covered work with a work licensed under version 3 of the GNU General Public License into a single combined work, and to convey the resulting work. The terms of this License will continue to apply to the part which is the covered work, but the work with which it is combined will remain governed by version 3 of the GNU General Public License. 14. Revised Versions of this License. The Free Software Foundation may publish revised and/or new versions of the GNU Affero General Public License from time to time. Such new versions will be similar in spirit to the present version, but may differ in detail to address new problems or concerns. Each version is given a distinguishing version number. If the Program specifies that a certain numbered version of the GNU Affero General Public License "or any later version" applies to it, you have the option of following the terms and conditions either of that numbered version or of any later version published by the Free Software Foundation. If the Program does not specify a version number of the GNU Affero General Public License, you may choose any version ever published by the Free Software Foundation. If the Program specifies that a proxy can decide which future versions of the GNU Affero General Public License can be used, that proxy's public statement of acceptance of a version permanently authorizes you to choose that version for the Program. Later license versions may give you additional or different permissions. However, no additional obligations are imposed on any author or copyright holder as a result of your choosing to follow a later version. 15. Disclaimer of Warranty. THERE IS NO WARRANTY FOR THE PROGRAM, TO THE EXTENT PERMITTED BY APPLICABLE LAW. EXCEPT WHEN OTHERWISE STATED IN WRITING THE COPYRIGHT HOLDERS AND/OR OTHER PARTIES PROVIDE THE PROGRAM "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESSED OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE. THE ENTIRE RISK AS TO THE QUALITY AND PERFORMANCE OF THE PROGRAM IS WITH YOU. SHOULD THE PROGRAM PROVE DEFECTIVE, YOU ASSUME THE COST OF ALL NECESSARY SERVICING, REPAIR OR CORRECTION. 16. Limitation of Liability. IN NO EVENT UNLESS REQUIRED BY APPLICABLE LAW OR AGREED TO IN WRITING WILL ANY COPYRIGHT HOLDER, OR ANY OTHER PARTY WHO MODIFIES AND/OR CONVEYS THE PROGRAM AS PERMITTED ABOVE, BE LIABLE TO YOU FOR DAMAGES, INCLUDING ANY GENERAL, SPECIAL, INCIDENTAL OR CONSEQUENTIAL DAMAGES ARISING OUT OF THE USE OR INABILITY TO USE THE PROGRAM (INCLUDING BUT NOT LIMITED TO LOSS OF DATA OR DATA BEING RENDERED INACCURATE OR LOSSES SUSTAINED BY YOU OR THIRD PARTIES OR A FAILURE OF THE PROGRAM TO OPERATE WITH ANY OTHER PROGRAMS), EVEN IF SUCH HOLDER OR OTHER PARTY HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH DAMAGES. 17. Interpretation of Sections 15 and 16. If the disclaimer of warranty and limitation of liability provided above cannot be given local legal effect according to their terms, reviewing courts shall apply local law that most closely approximates an absolute waiver of all civil liability in connection with the Program, unless a warranty or assumption of liability accompanies a copy of the Program in return for a fee. END OF TERMS AND CONDITIONS How to Apply These Terms to Your New Programs If you develop a new program, and you want it to be of the greatest possible use to the public, the best way to achieve this is to make it free software which everyone can redistribute and change under these terms. To do so, attach the following notices to the program. It is safest to attach them to the start of each source file to most effectively state the exclusion of warranty; and each file should have at least the "copyright" line and a pointer to where the full notice is found. <one line to give the program's name and a brief idea of what it does.> Copyright (C) <year> <name of author> This program is free software: you can redistribute it and/or modify it under the terms of the GNU Affero General Public License as published by the Free Software Foundation, either version 3 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Affero General Public License for more details. You should have received a copy of the GNU Affero General Public License along with this program. If not, see <https://www.gnu.org/licenses/>. Also add information on how to contact you by electronic and paper mail. If your software can interact with users remotely through a computer network, you should also make sure that it provides a way for users to get its source. For example, if your program is a web application, its interface could display a "Source" link that leads users to an archive of the code. There are many ways you could offer source, and different solutions will be better for different programs; see section 13 for the specific requirements. You should also get your employer (if you work as a programmer) or school, if any, to sign a "copyright disclaimer" for the program, if necessary. For more information on this, and how to apply and follow the GNU AGPL, see <https://www.gnu.org/licenses/>.
code-navigation, codebase, indexer, mcp, structural-analysis
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: GNU Affero General Public License v3", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Topic :: Software Development :: Libraries" ]
[]
null
null
>=3.11
[]
[]
[]
[ "pytest>=8.0; extra == \"dev\"", "ruff>=0.5; extra == \"dev\"", "mcp>=1.0; extra == \"mcp\"" ]
[]
[]
[]
[ "Homepage, https://github.com/MikeRecognex/mcp-codebase-index", "Repository, https://github.com/MikeRecognex/mcp-codebase-index" ]
twine/6.2.0 CPython/3.13.2
2026-02-20T22:04:31.334310
mcp_codebase_index-0.4.5.tar.gz
93,376
94/c4/20cf149d0b72e062c73821a24d28ca2c903737025143eb1a56bfa5ca7493/mcp_codebase_index-0.4.5.tar.gz
source
sdist
null
false
ab84bd689d63369baf0345ca0dc93f25
023494fc265a62e5ff60a29cf2b112a65c26d7fdb1ea48cf16db34fe0362333d
94c420cf149d0b72e062c73821a24d28ca2c903737025143eb1a56bfa5ca7493
null
[ "LICENSE" ]
204
2.4
azure-ai-agentserver-core
1.0.0b13
Agents server adapter for Azure AI
# Azure AI Agent Server Adapter for Python ## Getting started ```bash pip install azure-ai-agentserver-core ``` ## Key concepts This is the core package for Azure AI Agent server. It hosts your agent as a container on the cloud. You can talk to your agent using azure-ai-project sdk. ## Examples If your agent is not built using a supported framework such as LangGraph and Agent-framework, you can still make it compatible with Microsoft AI Foundry by manually implementing the predefined interface. ```python import datetime from azure.ai.agentserver.core import FoundryCBAgent from azure.ai.agentserver.core.models import ( CreateResponse, Response as OpenAIResponse, ) from azure.ai.agentserver.core.models.projects import ( ItemContentOutputText, ResponsesAssistantMessageItemResource, ResponseTextDeltaEvent, ResponseTextDoneEvent, ) def stream_events(text: str): assembled = "" for i, token in enumerate(text.split(" ")): piece = token if i == len(text.split(" ")) - 1 else token + " " assembled += piece yield ResponseTextDeltaEvent(delta=piece) # Done with text yield ResponseTextDoneEvent(text=assembled) async def agent_run(request_body: CreateResponse): agent = request_body.agent print(f"agent:{agent}") if request_body.stream: return stream_events("I am mock agent with no intelligence in stream mode.") # Build assistant output content output_content = [ ItemContentOutputText( text="I am mock agent with no intelligence.", annotations=[], ) ] response = OpenAIResponse( metadata={}, temperature=0.0, top_p=0.0, user="me", id="id", created_at=datetime.datetime.now(), output=[ ResponsesAssistantMessageItemResource( status="completed", content=output_content, ) ], ) return response my_agent = FoundryCBAgent() my_agent.agent_run = agent_run if __name__ == "__main__": my_agent.run() ``` ## Troubleshooting First run your agent with azure-ai-agentserver-core locally. If it works on local by failed on cloud. Check your logs in the application insight connected to your Azure AI Foundry Project. ### Reporting issues To report an issue with the client library, or request additional features, please open a GitHub issue [here](https://github.com/Azure/azure-sdk-for-python/issues). Mention the package name "azure-ai-agents" in the title or content. ## Next steps Please visit [Samples](https://github.com/Azure/azure-sdk-for-python/tree/main/sdk/agentserver/azure-ai-agentserver-core/samples) folder. There are several cases for you to build your agent with azure-ai-agentserver ## Contributing This project welcomes contributions and suggestions. Most contributions require you to agree to a Contributor License Agreement (CLA) declaring that you have the right to, and actually do, grant us the rights to use your contribution. For details, visit https://cla.microsoft.com. When you submit a pull request, a CLA-bot will automatically determine whether you need to provide a CLA and decorate the PR appropriately (e.g., label, comment). Simply follow the instructions provided by the bot. You will only need to do this once across all repos using our CLA. This project has adopted the [Microsoft Open Source Code of Conduct][code_of_conduct]. For more information, see the Code of Conduct FAQ or contact opencode@microsoft.com with any additional questions or comments.
text/markdown
null
Microsoft Corporation <azpysdkhelp@microsoft.com> License-Expression: MIT
null
null
null
azure, azure sdk
[ "Development Status :: 4 - Beta", "Programming Language :: Python", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13" ]
[]
null
null
>=3.10
[]
[]
[]
[ "azure-monitor-opentelemetry<1.8.5,>=1.5.0", "azure-ai-projects>=2.0.0b1", "azure-ai-agents==1.2.0b5", "azure-core>=1.35.0", "azure-identity>=1.25.1", "openai>=1.80.0", "opentelemetry-api>=1.35", "opentelemetry-exporter-otlp-proto-http", "starlette>=0.45.0", "uvicorn>=0.31.0", "aiohttp>=3.13.0", "cachetools>=6.0.0" ]
[]
[]
[]
[ "repository, https://github.com/Azure/azure-sdk-for-python" ]
RestSharp/106.13.0.0
2026-02-20T22:04:12.320020
azure_ai_agentserver_core-1.0.0b13-py3-none-any.whl
168,393
12/08/eb2c11a5415a1303fc8f7cb583809bcd8431d9c7b704520bd0975dc17a2f/azure_ai_agentserver_core-1.0.0b13-py3-none-any.whl
py3
bdist_wheel
null
false
33c448084e816213af93557c40620384
994ad8c4205c5b3d5f98d17beaf6bea3e14aacde748530979af9a4e6a65ce8c0
1208eb2c11a5415a1303fc8f7cb583809bcd8431d9c7b704520bd0975dc17a2f
null
[]
305
2.4
nettracer3d
1.4.7
GUI for intializing and analyzing networks from segmentations of three dimensional images.
NetTracer3D is a python package developed for both 2D and 3D analysis of microscopic images in the .tif file format. It supports generation of 3D networks showing the relationships between objects (or nodes) in three dimensional space, either based on their own proximity or connectivity via connecting objects such as nerves or blood vessels. In addition to these functionalities are several advanced 3D data processing algorithms, such as labeling of branched structures or abstraction of branched structures into networks. Note that nettracer3d uses segmented data, which can be segmented from other softwares such as ImageJ and imported into NetTracer3D, although it does offer its own segmentation via intensity and volumetric thresholding, or random forest machine learning segmentation. NetTracer3D currently has a fully functional GUI. To use the GUI, after installing the nettracer3d package via pip, enter the command 'nettracer3d' in your command prompt: --- Documentation --- Please see: https://nettracer3d.readthedocs.io/en/latest/ --- Video Tutorial --- Please see: https://www.youtube.com/watch?v=_4uDy0mzG94&list=PLsrhxiimzKJMZ3_gTWkfrcAdJQQobUhj7 --- Installing as a Python package --- 1. **Get Python and Pip on your path**: To install nettracer3d, first install Python version 3.12. Make sure the Python installation installs pip, and that both Python and pip are available on your PATH. I recommend installing Python using the installer which is available here. Make sure to check the option to 'add Python to PATH' when it appears: https://www.python.org/downloads/ 2. **Base Package**: Next, use this command in your command terminal * pip install nettracer3d 3. **For 3D Displays**: Or if you also want Napari for 3D displays: * pip install nettracer3d[viz] 4. **Optional Performance Boost**: If you are trying to process large images, you may also want to include the 'edt' module in your package. This will allow parallelized CPU calculations for several of the search functions which can increase their speed by an order of magnitude or more depending on how many cores your CPU has. This can be a major benefit if you have a strong CPU and sufficient RAM. It requires an extra pre-installation step, thus is not included by default. You will also have to install the C++ build tools from windows. Please head to this link, then download and run the installer: https://visualstudio.microsoft.com/visual-cpp-build-tools/. In the menu of the installer, select the 'Desktop Development with C++' option, then proceed to download/install it using the installation menu. You will likely want to be using the Python distributed from the actual Python website and not the windows store (or elsewhere) or the edt module may not work properly. To bundle with edt use: * pip install nettracer3d[edt] 5. **Recommended full package**: Or if you want to just get both edt and napari at once: * pip install nettracer3d[rec] 6. Likewise, if you already installed the default version, you can add napari and/or edt with just: * pip install edt * pip install napari --- Installing as a Python package in Anaconda--- I recommend installing the program as an Anaconda package to ensure its modules are work together on your specific system: (Install anaconda at the link below, set up a new python env for nettracer3d, then use the same pip command). https://www.anaconda.com/download?utm_source=anacondadocs&utm_medium=documentation&utm_campaign=download&utm_content=installwindows --- Using the downloadable version --- Alternatively, you can download a compiled .exe of version 1.2.7 here: https://doi.org/10.5281/zenodo.17873800 Unzip the folder, then double click the NetTracer3D executable to run the program. Note that this version will be missing a few features compared to the Python package, namely GPU segmentation support and the ability to print updates to the command window. It will also not be updated as often. --- Optional Packages --- I recommend including Napari (Chi-Li Chiu, Nathan Clack, the napari community, napari: a Python Multi-Dimensional Image Viewer Platform for the Research Community, Microscopy and Microanalysis, Volume 28, Issue S1, 1 August 2022, Pages 1576–1577, https://doi.org/10.1017/S1431927622006328) in the download as well, which allows NetTracer3D to use 3D displays. The standard package only comes with its native 2D slice display window. If Napari is present, all 3D images and overlays from NetTracer3D can be easily displayed in 3D with a click of a button. To package with Napari, use this install command instead: pip install nettracer3d[viz] Additionally, for easy access to high-quality cell segmentation, as of version 0.8.2, NetTracer3D can be optionally packaged with Cellpose3. (Stringer, C., Pachitariu, M. Cellpose3: one-click image restoration for improved cellular segmentation. Nat Methods 22, 592–599 (2025). https://doi.org/10.1038/s41592-025-02595-5) Cellpose3 is not involved with the rest of the program in any way, although its GUI can be opened from NetTracer3D's GUI, provided both are installed in the same environment. It is a top-tier cell segmenter which can assist in the production of cell networks. To include Cellpose3 in the install, use this command: pip install nettracer3d[cellpose] Alternatively, Napari, Cellpose, and edt can be included in the package with this command: (Or they can be independently installed with pip from the base package env) pip install nettracer3d[all] --- GPU --- NetTracer3D is mostly CPU-bound, but a few functions can optionally use the GPU. To install optional GPU functionalities, first set up a CUDA toolkit that runs with the GPU on your machine. This requires an NVIDIA GPU. Then, find your GPUs compatible CUDA toolkit and install it with the auto-installer from the NVIDIA website: https://developer.nvidia.com/cuda-toolkit With a CUDA toolkit installed, use: pip install nettracer3d[CUDA11] #If your CUDA toolkit is version 11 pip install nettracer3d[CUDA12] #If your CUDA toolkit is version 12 pip install nettracer3d[cupy] #For the generic cupy library (The above two are usually the ones you want) Or if you've already installed the NetTracer3D base package and want to get just the GPU associated packages: pip install cupy-cuda11x #If your CUDA toolkit is version 11 pip install cupy-cuda12x #If your CUDA toolkit is version 12 pip install cupy #For the generic cupy library (The above two are usually the ones you want) While not related to NetTracer3D, if you want to use Cellpose3 (for which GPU-usage is somewhat obligatory) to help segment cells for any networks, you will also want to install pytorch here: https://pytorch.org/. Use the pytorch build menu on this webpage to find a pip install command that is compatible with Python and your CUDA version. This gui is built from the PyQt6 package and therefore may not function on dockers or virtual envs that are unable to support PyQt6 displays. NetTracer3D is freely available for academic and nonprofit use and can obtained from pip (pip install nettracer3d), provided that citation is included in any abstract, paper, or presentation utilizing NetTracer3D. (The official paper to cite is coming soon) NetTracer3D was developed by Liam McLaughlin while working under Dr. Sanjay Jain at Washington University School of Medicine. -- Version 1.4.7 Updates -- * Most of the UMAP outputs are now interactable - select groups of nodes of interest (linked to the main image viewer) and flexibly configure identity vs community rendering. It's also possible to now save the UMAP embedding schema which is the big computational hurdle so you can calculate a big task and load it faster later. * The identity render-ers that encode identities as colored overlays no longer make use of 'multi-identity' nodes as their own category. There were often so many in some cases that this was not visually useful so I pulled it. * For the violin plots you can know flexibly input how many channels you want to show.
text/markdown
null
Liam McLaughlin <liamm@wustl.edu>
null
null
null
null
[ "Programming Language :: Python :: 3", "License :: Other/Proprietary License", "Operating System :: OS Independent" ]
[]
null
null
>=3.7
[]
[]
[]
[ "numpy", "scipy", "scikit-image", "Pillow", "matplotlib", "networkx", "opencv-python-headless", "openpyxl", "pandas", "tifffile", "qtrangeslider", "PyQt6", "pyqtgraph", "scikit-learn", "setuptools", "umap-learn", "numba", "cupy-cuda11x; extra == \"cuda11\"", "cupy-cuda12x; extra == \"cuda12\"", "cupy; extra == \"cupy\"", "cellpose[GUI]; extra == \"cellpose\"", "napari; extra == \"viz\"", "napari; extra == \"rec\"", "edt; extra == \"rec\"", "igraph; extra == \"rec\"", "leidenalg; extra == \"rec\"", "edt; extra == \"edt\"", "cellpose[GUI]; extra == \"all\"", "napari; extra == \"all\"", "edt; extra == \"all\"" ]
[]
[]
[]
[ "Documentation, https://nettracer3d.readthedocs.io/en/latest/", "Youtube_Tutorial, https://www.youtube.com/watch?v=_4uDy0mzG94&list=PLsrhxiimzKJMZ3_gTWkfrcAdJQQobUhj7", "Downloadable_Version, https://doi.org/10.5281/zenodo.17873800" ]
twine/6.1.0 CPython/3.13.2
2026-02-20T22:04:09.073937
nettracer3d-1.4.7.tar.gz
442,402
eb/30/267da2bba68fa5a455199594e3cdf440777ecfde0c3afc6c010fa3dbe975/nettracer3d-1.4.7.tar.gz
source
sdist
null
false
5fab8bd9ff6c2f7f80a84caeb2488dbe
7ae832ec1ae34a52eb627676b791601ff7124347cae7ad860cd72c5fa9918648
eb30267da2bba68fa5a455199594e3cdf440777ecfde0c3afc6c010fa3dbe975
null
[ "LICENSE" ]
222
2.4
pyhw
0.16.5
PyHw, a neofetch-like command line tool for fetching system information but written mostly in python.
# PyHw, a neofetch-like system information fetching tool [![Downloads](https://static.pepy.tech/badge/pyhw)](https://pepy.tech/project/pyhw) ![PyPI - Version](https://img.shields.io/pypi/v/pyhw?label=version) ![Static Badge](https://img.shields.io/badge/Python-3.9%2B-green) ![Static Badge](https://img.shields.io/badge/macOS-11%2B-green) ![Static Badge](https://img.shields.io/badge/Linux-blue) ![Static Badge](https://img.shields.io/badge/FreeBSD-red) ![Static Badge](https://img.shields.io/badge/Windows-yellow) ![Static Badge](https://img.shields.io/badge/amd64-green) ![Static Badge](https://img.shields.io/badge/aarch64-blue) ![Static Badge](https://img.shields.io/badge/arm32-yellow) ![Static Badge](https://img.shields.io/badge/riscv64-%238A2BE2) ![Static Badge](https://img.shields.io/badge/ppc64-orange) ![Static Badge](https://img.shields.io/badge/s390x-red) PyHw, a neofetch-like command line tool for fetching system information but written mostly in Python. This project is a Python reimplementation of [neofetch](https://github.com/dylanaraps/neofetch) and references the [fastfetch](https://github.com/fastfetch-cli/fastfetch) project for logo style settings. Since this project is implemented in Python, it will be easier to maintain and extend than bash and c implementation. Also, this project only relies on the Python standard library, so you can run it on any device that has a Python environment (I hope so 🤔). [//]: # (![demo]&#40;https://i2.imgs.ovh/d/BQACAgUAAx0EUvSR8wACMvpmyFVohzKxLcUdLiJaEa3wlo_OrQACuw4AAoX-QVaSpG0-rTAeRTUE&#41;) [//]: # (![demo]&#40;https://files.catbox.moe/xx58xy.jpg&#41;) [//]: # (![demo]&#40;https://files.catbox.moe/2d21fu.jpg&#41;) ![demo](https://files.catbox.moe/ik4ioi.png) - [1. Install](#1-install) - [2. Usability](#2-usability) - [3. Add Logo](#3-add-logo) - [4. Build from source](#4-build-from-source) - [5. Test Package](#5-test-package) - [6. Troubleshooting](#6-troubleshooting) ## 1. Install There are already a lot of similar tools so you can choose any of them; they're all essentially no different. If you want to try this tool, There are two convenient ways to install it. ### 1.1 Install by pipx **pipx** is an amazing tool to help you install and run applications written in Python. It is more like **brew** or **apt**. You can find more information about it here [pipx](https://github.com/pypa/pipx). **pipx** is available on almost all major platforms and is usually provided by the corresponding package manager. If you haven't used pipx before, you can refer to this [document](https://pipx.pypa.io/stable/installation/) to install it. You can install pyhw by the following command: ```shell pipx install pyhw ``` You can then use this tool directly from the command line with the following command, just like neofetch. ```shell pyhw ``` ### 1.2 Install by pip In any case, pip is always available, so if you can't install this program using **pipx**, you can install pyhw by the following command: ```shell pip install pyhw ``` To upgrade pyhw: ```shell pip install pyhw -U # or pip install pyhw --upgrade ``` You can then use this tool directly from the command line with the following command, just like neofetch. ```shell pyhw # or python -m pyhw ``` Please note that the command line entry for __pyhw__ is created by pip, and depending on the user, this entry may not in the __system PATH__. If you encounter this problem, pip will give you a prompt, follow the prompts to add entry to the __system PATH__. ## 2. Usability ### Tested Platform The following platforms have been tested and are known to work with this package: * macOS: arm64, x86_64 * Linux: arm64, x86_64, riscv64, ppc64le, mips64el, s390x * FreeBSD: arm64, x86_64 * Windows 10: x86_64 * Windows 11: arm64, x86_64 For more detailed information, please refer to [Tested Platform](docs/tested_platform.md). Please note that this package requires `Python 3.9`, so very old versions of Linux may not be supported. ### Features The functionality of this package varies slightly on different operating systems and architectures, please refer to [this](docs/functionality.md) documentation for details. ## 3. Add Logo 1. Create a file named **\<os>.pyhw** in **logo/ascii** folder 2. Modify **colorConfig.py** file to add a new logo style 3. Update **pyhwUtil.py** to enable new logo style. 4. You may create a new `PR` to add your logo style to the main repository. ## 4. Build from source ### 4.1 Dependencies This package was originally implemented in pure python and only depends on the python standard library. However, in subsequent development, the code for the pci part was separated into a separate package **pypci-ng**, which can be obtained using pip (or check out [this](https://github.com/xiaoran007/pypci) GitHub repository). ### 4.2 Build tools Make sure the following Python build tools are already installed. * setuptools * build * twine Newer versions of twine requires the following dependencies are up to date: * setuptools * build * twine * packaging ### 4.3 Build package clone the project, and run: ```shell python -m build ``` After the build process, the source package and the binary whl package can be found in the dist folder. Then you can use the following command to install the new package. ```shell pip install dist/*.whl --force-reinstall ``` ### 4.4 Build Full Feature package Currently, build process relay on swiftc and macOS IOKit framework. To build Full Feature Package from source, you need a Mac machine with macOS 11 and newer. Simply type: ```shell make build make install ``` ## 5. Test Package If you have docker installed, you can test this package through docker by type: ```shell make test # local build make test-pypi # release version ``` ## 6. Troubleshooting ### 6.1 Important note about debian 12: If you use system pip to install pyhw, you will encounter this problem on debian12 and some related distributions (like Ubuntu 24.04): ```text error: externally-managed-environment × This environment is externally managed ╰─> To install Python packages system-wide, try apt install python3-xyz, where xyz is the package you are trying to install. If you wish to install a non-Debian-packaged Python package, create a virtual environment using python3 -m venv path/to/venv. Then use path/to/venv/bin/python and path/to/venv/bin/pip. Make sure you have python3-full installed. For more information visit http://rptl.io/venv note: If you believe this is a mistake, please contact your Python installation or OS distribution provider. You can override this, at the risk of breaking your Python installation or OS, by passing --break-system-packages. hint: See PEP 668 for the detailed specification. ``` This is due to the fact that system python is not supposed to be managed by pip. You can simply use **pipx** to install **pyhw**. Or you can use a virtual environment (venv), conda environment or force remove this restriction (not recommended).
text/markdown
Xiao Ran
null
null
Xiao Ran <xiaoran.007@icloud.com>
null
neofetch, system information, command line tool, python, hardware information, fastfetch, fetching
[ "Development Status :: 5 - Production/Stable", "Programming Language :: Python :: 3", "Operating System :: OS Independent" ]
[]
null
null
>=3.9
[]
[]
[]
[ "pypci-ng>=0.3.4" ]
[]
[]
[]
[ "homepage, https://github.com/xiaoran007/pyhw" ]
twine/6.2.0 CPython/3.9.24
2026-02-20T22:03:46.158648
pyhw-0.16.5.tar.gz
172,741
54/d7/c98f31b698028f2317529420beff120721216b68740a7932eaa762320b91/pyhw-0.16.5.tar.gz
source
sdist
null
false
a5df945c9d830d3ba8ba0357b936f078
a78337f8adafca3931884a784fcf1774485d9b165724f07f67a2cdf8af4100d6
54d7c98f31b698028f2317529420beff120721216b68740a7932eaa762320b91
BSD-3-Clause
[ "LICENSE" ]
206
2.4
ispider
0.8.6
A high-speed web spider for massive scraping.
# ispider_core **ispider** is a module to spider websites - Multicore and multithreaded - Accepts hundreds/thousands of websites/domains as input - Sparse requests to avoid repeated calls against the same domain - The `httpx` engine works in asyncio blocks defined by `settings.ASYNC_BLOCK_SIZE`, so total concurrent threads are `ASYNC_BLOCK_SIZE * POOLS` - It supports retry with different engines (httpx, curl, seleniumbase [testing]) It was designed for maximum speed, so it has some limitations: - As of v0.7, it does not support files (pdf, video, images, etc); it only processes HTML # HOW IT WORKS - SIMPLE **-- Crawl - Depth == 0** - Get all the landing pages for domains in the provided list. - If "robots" is selected, download the `robots.txt` file. - If "sitemaps" is selected, parse the `robots.txt` and retrieve all the sitemaps. - All data is saved under `USER_DATA/data/dumps/dom_tld`. **-- Spider - Depth > 0** - Extract all links from landing pages and sitemaps. - Download the HTML pages, extract internal links, and follow them recursively. # HOW IT WORKS - MORE DETAILED #### Crawl - Depth == 0 - Create objects in the form (`('https://domain.com', 'landing_page', 'domain.com', depth, retries, engine)`) - Add them to the LIFO queue `qout` - A thread retrieves elements from `qout` in variable-size blocks (depending on `QUEUE_MAX_SIZE`) - Fill a FIFO queue `qin` - Different workers (defined in `settings.POOLS`) get elements from `qin` and download them to `USER_DATA/data/dumps/dom_tld` - Landing pages are saved as `_.html` - Each worker processes the landing page; if the result is OK (`status_code == 200`), it tries to get `robots.txt` - On failure, it tries the next available engine (fallback) - It creates an object (`('https://domain.com/robots.txt', 'robots', 'domain.com', depth=1, retries=0, engine)`) - Each worker retrieves the `robots.txt`; if `"sitemaps"` is defined in `settings.CRAWL_METHODS`, it attempts to get all sitemaps from `robots.txt` and `dom_tld/sitemaps.xml` - It creates objects (`('https://domain.com/sitemap.xml', 'sitemaps', 'domain.com', depth=1, retries=0, engine)`) and for other sitemaps found in `robots.txt` - Every successful or failed download is logged as a row in `USER_FOLDER/jsons/crawl_conn_meta*json` with all information available from the engine; these files are useful for statistics/reports from the spider - When there are no more elements in `qin`, after a 90-second timeout, jobs stop. #### Spider - Depths > 0 - It reads entries from `USER_FOLDER/jsons/crawl_conn_meta*json` for the domains in the list - It retrieves landing pages and sitemaps - If sitemaps are compressed, it uncompresses them - Extract all links from landing pages and sitemaps - Create objects (`('https://domain.com/link1', 'internals', 'domain.com', depth=2, retries=0, engine)`) - Use the same engine that was used for the last successful request to the domain TLD - Add these objects to `qout` - Thread `qin` moves blocks from `qout` to `qin`, sparsing them - Download all links, save them, and save data in JSON - Parse the HTML, extract all INTERNAL links, follow them recursively, increasing depth #### Schema This is the projectual schema of the crawler/spider ![alt text](https://i.imgur.com/vA05tbF.png) # USAGE Install it ``` pip install ispider ``` First use ``` from ispider_core import ISpider if __name__ == '__main__': # Check the readme for the complete avail parameters config_overrides = { 'USER_FOLDER': '/Your/Dump/Folder', 'POOLS': 64, 'ASYNC_BLOCK_SIZE': 32, 'MAXIMUM_RETRIES': 2, 'CRAWL_METHODS': [], 'CODES_TO_RETRY': [430, 503, 500, 429], 'CURL_INSECURE': True, 'ENGINES': ['curl'], 'EXCLUDED_DOMAINS': ['facebook.com', 'instagram.com'] } # Specify a list of domains doms = ['domain1.com', 'domain2.com'....] # Run with ISpider(domains=doms, **config_overrides) as spider: spider.run() ``` # TO KNOW At first execution, - It creates the folder settings.USER_FOLDER - It creates settings.USER_FOLDER/data/ with dumps/ and jsons/ - settings.USER_FOLDER/data/dumps are the downloaded websites - settings.USER_FOLDER/data/jsons are the connection results for every request # SETTINGS Actual default settings are: """ ## ********************************* ## GENERIC SETTINGS # Output folder for controllers, dumps and jsons USER_FOLDER = "~/.ispider/" # Log level LOG_LEVEL = 'DEBUG' ## i.e., status_code = 430 CODES_TO_RETRY = [430, 503, 500, 429] MAXIMUM_RETRIES = 2 # Delay time after some status code to be retried TIME_DELAY_RETRY = 0 ## Number of concurrent connection on the same process during crawling # Concurrent por process ASYNC_BLOCK_SIZE = 4 # Concurrent processes (number of cores used, check your CPU spec) POOLS = 4 # Max timeout for connecting, TIMEOUT = 5 # This need to be a list, # curl is used as subprocess, so be sure you installed it on your system # Retry will use next available engine. # The script begins wit the suprfast httpx # If fail, try with curl # If fail, it tries with seleniumbase, headless and uc mode activate ENGINES = ['httpx', 'curl', 'seleniumbase'] CURL_INSECURE = False ## ********************************* # CRAWLER # File size # Max file size dumped on the disk. # This to avoid big sitemaps with errors. MAX_CRAWL_DUMP_SIZE = 52428800 # Max depth to follow in sitemaps SITEMAPS_MAX_DEPTH = 2 # Crawler will get robots and sitemaps too CRAWL_METHODS = ['robots', 'sitemaps'] ## ********************************* ## SPIDER # Queue max, till 1 billion is ok on normal systems QUEUE_MAX_SIZE = 100000 # Max depth to follow in websites WEBSITES_MAX_DEPTH = 2 # This is not implemented yet MAX_PAGES_POR_DOMAIN = 1000000 # This try to exclude some kind of files # It also test first bits of content of some common files, # to exclude them even if online element has no extension EXCLUDED_EXTENSIONS = [ "pdf", "csv", "mp3", "jpg", "jpeg", "png", "gif", "bmp", "tiff", "webp", "svg", "ico", "tif", "jfif", "eps", "raw", "cr2", "nef", "orf", "arw", "rw2", "sr2", "dng", "heif", "avif", "jp2", "jpx", "wdp", "hdp", "psd", "ai", "cdr", "ppsx" "ics", "ogv", "mpg", "mp4", "mov", "m4v", "zip", "rar" ] # Exclude all urls that contains this REGEX EXCLUDED_EXPRESSIONS_URL = [ # r'test', ] # If not empty, follow only URLs that match these regex patterns INCLUDED_EXPRESSIONS_URL = [ # r'/\d{4}/\d{2}/\d{2}/', ] # Exclude specific domains from crawling/spidering. # Accepts values like "example.com" or full URLs. EXCLUDED_DOMAINS = [] """ # NOTES - Deduplication is not 100% safe, sometimes pages are downloaded multiple times, and skipped in file check. On ~10 domains, check duplication has small delay. But on 10000 domains after 500k links, the domain list is so big that checking if a link is already downloaded or not was decreasing considerably the speed (from 30000 urls/min to 300 urls/min). That's why I preferred avoid a list, and left just "check file". ## SEO checks (modular) You can run independent SEO checks during crawling/spidering. Results are stored in each JSON response row under `seo_issues`. Available checks: - `response_crawlability`: flags 3xx/4xx/5xx, redirect chains, and timeouts. - `broken_links`: generic status >= 400 detector. - `http_status_503`: dedicated 503 detector. - `title_meta_quality`: validates `<title>` and meta description length/presence and flags `title == h1`. - `h1_too_long`: validates H1 length threshold. - `heading_structure`: checks h1 count and heading-order skips. - `indexability_canonical`: checks canonical presence/self-reference, homepage canonicals, and `noindex` directives. - `schema_news_article`: detects `NewsArticle` structured data and required properties. - `image_optimization`: flags missing image dimensions/ALT and oversized hero hints. - `internal_linking`: flags weak anchors, no internal links, and too many external links. - `url_hygiene`: validates URL length/case/params/special chars and the newsroom pattern `/yyyy/mm/dd/slug/`. - `content_length`: flags thin content (default `<250` words). - `security_headers`: checks HSTS, CSP, and X-Frame-Options. ### SEO issue codes (priority + short description) | Code | Priority | Description | |---|---|---| | `BROKEN_LINK` | medium | URL returned an HTTP status code >= 400. | | `CANONICAL_MISSING` | medium | Canonical tag is missing. | | `CANONICAL_NOT_SELF` | low | Canonical URL is not self-referential. | | `CANONICAL_TO_HOMEPAGE` | high | Canonical points to homepage from an internal page. | | `CONTENT_TOO_THIN` | medium | Visible content word count is below the configured minimum. | | `H1_MISSING` | high | No H1 heading found on the page. | | `H1_MULTIPLE` | high | More than one H1 heading found. | | `H1_TOO_LONG` | low | H1 text length exceeds configured maximum (`SEO_H1_MAX_CHARS`). | | `HEADING_ORDER_SKIP` | low | Heading hierarchy skips levels (for example `h2` -> `h4`). | | `HERO_IMAGE_FETCHPRIORITY_MISSING` | low | First image is missing `fetchpriority=high`. | | `HERO_IMAGE_TOO_LARGE` | medium | Hero image appears larger than configured size threshold. | | `HTTP_3XX` | low | Response is a redirect (3xx). | | `HTTP_4XX` | high | Response is a client error (4xx). | | `HTTP_5XX` | high | Response is a server error (5xx). | | `HTTP_503` | high | Response specifically returned 503 Service Unavailable. | | `IMAGE_ALT_MISSING` | low | At least one image is missing ALT text. | | `IMAGE_LAZY_LOADING_MISSING` | low | Non-hero image missing `loading=lazy`. | | `META_DESCRIPTION_LENGTH` | low | Meta description length is outside recommended range. | | `META_DESCRIPTION_MISSING` | medium | Meta description is missing. | | `NOINDEX_DETECTED` | high | `noindex` detected in meta robots or x-robots-tag. | | `NO_INTERNAL_LINKS` | medium | No internal links found on the page. | | `REDIRECT_CHAIN` | medium | Redirect chain length is greater than 1. | | `REQUEST_TIMEOUT` | high | Request timed out. | | `SCHEMA_NEWSARTICLE_MISSING` | high | `NewsArticle` JSON-LD schema not found. | | `SCHEMA_REQUIRED_FIELDS_MISSING` | high | `NewsArticle` schema is missing required fields. | | `SECURITY_HEADERS_MISSING` | low | One or more security headers are missing (HSTS, CSP, X-Frame-Options). | | `TITLE_EQUALS_H1` | low | `<title>` is identical to H1. | | `TITLE_LENGTH` | medium | `<title>` length is outside recommended range. | | `TITLE_MISSING` | high | `<title>` tag is missing. | | `TOO_MANY_EXTERNAL_LINKS` | low | Unique external domains exceed configured threshold. | | `URL_HAS_PARAMETERS` | low | URL contains query parameters. | | `URL_NEWS_PATTERN_MISMATCH` | medium | URL does not match expected `/yyyy/mm/dd/slug/` pattern. | | `URL_SPECIAL_CHARS` | low | URL path contains special characters. | | `URL_TOO_LONG` | low | URL length exceeds configured threshold. | | `URL_UPPERCASE` | low | URL path contains uppercase letters. | | `WEAK_ANCHOR_TEXT` | low | Generic anchor texts detected (for example “read more”, “click here”). | Configure with settings: ```python config_overrides = { 'SEO_CHECKS_ENABLED': True, 'SEO_ENABLED_CHECKS': ['response_crawlability', 'title_meta_quality', 'schema_news_article'], 'SEO_DISABLED_CHECKS': ['http_status_503'], 'SEO_H1_MAX_CHARS': 70, } ``` Tip for Google News-focused runs: combine `INCLUDED_EXPRESSIONS_URL` with a day filter (example: `r'^.*/2026/02/07/.*$'`) and keep `response_crawlability`, `indexability_canonical`, and `schema_news_article` enabled. To add a new check, create a class in `ispider_core/seo/checks/` with `name` and `run(resp)` and register it in `ispider_core/seo/runner.py`.
text/markdown
null
Daniele Rugginenti <daniele.rugginenti@gmail.com>
null
null
null
null
[]
[]
null
null
null
[]
[]
[]
[ "aiohttp", "beautifulsoup4", "lxml", "tqdm", "requests", "seleniumbase", "httpx", "nslookup", "tldextract", "concurrent_log_handler", "colorlog", "brotli", "validators", "w3lib", "pybloom_live", "uvicorn", "fastapi", "pandas" ]
[]
[]
[]
[ "Homepage, https://github.com/danruggi/ispider" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T22:01:43.835147
ispider-0.8.6.tar.gz
50,137
38/40/1867a68387d3820866d573fad60673f2875657c8324a364e52348d6b88e9/ispider-0.8.6.tar.gz
source
sdist
null
false
e6f7908d173ca1a6cf7a5b4f0a07b5af
c47041767ca172fd2574ec7286fbaca21fb1cdaddc7effa5ab6e42693ee08a00
38401867a68387d3820866d573fad60673f2875657c8324a364e52348d6b88e9
MIT
[ "LICENCE" ]
202
2.3
fragment-py
0.6.0
The official Python library for the fragment API
# Fragment Python API library <!-- prettier-ignore --> [![PyPI version](https://img.shields.io/pypi/v/fragment-py.svg?label=pypi%20(stable))](https://pypi.org/project/fragment-py/) The Fragment Python library provides convenient access to the Fragment REST API from any Python 3.9+ application. The library includes type definitions for all request params and response fields, and offers both synchronous and asynchronous clients powered by [httpx](https://github.com/encode/httpx). It is generated with [Stainless](https://www.stainless.com/). ## Documentation The full API of this library can be found in [api.md](https://github.com/fragment-dev/fragment-py/tree/main/api.md). ## Installation ```sh # install from PyPI pip install fragment-py ``` ## Usage The full API of this library can be found in [api.md](https://github.com/fragment-dev/fragment-py/tree/main/api.md). ```python from fragment import Fragment client = Fragment() response = client.transactions.create_allocations( id="txn_abc123", allocation_updates=[ { "amount": "1000", "invoice_id": "inv_abc123", "op": "add", "type": "invoice_payin", "user": {"id": "user_abc123"}, } ], version=1, ) print(response.data) ``` ## Async usage Simply import `AsyncFragment` instead of `Fragment` and use `await` with each API call: ```python import asyncio from fragment import AsyncFragment client = AsyncFragment() async def main() -> None: response = await client.transactions.create_allocations( id="txn_abc123", allocation_updates=[ { "amount": "1000", "invoice_id": "inv_abc123", "op": "add", "type": "invoice_payin", "user": {"id": "user_abc123"}, } ], version=1, ) print(response.data) asyncio.run(main()) ``` Functionality between the synchronous and asynchronous clients is otherwise identical. ### With aiohttp By default, the async client uses `httpx` for HTTP requests. However, for improved concurrency performance you may also use `aiohttp` as the HTTP backend. You can enable this by installing `aiohttp`: ```sh # install from PyPI pip install fragment-py[aiohttp] ``` Then you can enable it by instantiating the client with `http_client=DefaultAioHttpClient()`: ```python import asyncio from fragment import DefaultAioHttpClient from fragment import AsyncFragment async def main() -> None: async with AsyncFragment( http_client=DefaultAioHttpClient(), ) as client: response = await client.transactions.create_allocations( id="txn_abc123", allocation_updates=[ { "amount": "1000", "invoice_id": "inv_abc123", "op": "add", "type": "invoice_payin", "user": {"id": "user_abc123"}, } ], version=1, ) print(response.data) asyncio.run(main()) ``` ## Using types Nested request parameters are [TypedDicts](https://docs.python.org/3/library/typing.html#typing.TypedDict). Responses are [Pydantic models](https://docs.pydantic.dev) which also provide helper methods for things like: - Serializing back into JSON, `model.to_json()` - Converting to a dictionary, `model.to_dict()` Typed requests and responses provide autocomplete and documentation within your editor. If you would like to see type errors in VS Code to help catch bugs earlier, set `python.analysis.typeCheckingMode` to `basic`. from datetime import datetime ## Nested params Nested parameters are dictionaries, typed using `TypedDict`, for example: ```python from fragment import Fragment client = Fragment() transaction = client.transactions.create( account={}, allocations=[ { "amount": "1000", "invoice_id": "inv_abc123", "type": "invoice_payin", "user": {"id": "user_abc123"}, } ], amount="-1000", currency="USD", external_id="bank_txn_123", posted=datetime.fromisoformat("2026-02-12T00:00:00.000"), ) print(transaction.account) ``` ## Handling errors When the library is unable to connect to the API (for example, due to network connection problems or a timeout), a subclass of `fragment.APIConnectionError` is raised. When the API returns a non-success status code (that is, 4xx or 5xx response), a subclass of `fragment.APIStatusError` is raised, containing `status_code` and `response` properties. All errors inherit from `fragment.APIError`. ```python import fragment from fragment import Fragment client = Fragment() try: client.transactions.create_allocations( id="txn_abc123", allocation_updates=[ { "amount": "1000", "invoice_id": "inv_abc123", "op": "add", "type": "invoice_payin", "user": {"id": "user_abc123"}, } ], version=1, ) except fragment.APIConnectionError as e: print("The server could not be reached") print(e.__cause__) # an underlying Exception, likely raised within httpx. except fragment.RateLimitError as e: print("A 429 status code was received; we should back off a bit.") except fragment.APIStatusError as e: print("Another non-200-range status code was received") print(e.status_code) print(e.response) ``` Error codes are as follows: | Status Code | Error Type | | ----------- | -------------------------- | | 400 | `BadRequestError` | | 401 | `AuthenticationError` | | 403 | `PermissionDeniedError` | | 404 | `NotFoundError` | | 422 | `UnprocessableEntityError` | | 429 | `RateLimitError` | | >=500 | `InternalServerError` | | N/A | `APIConnectionError` | ### Retries Certain errors are automatically retried 2 times by default, with a short exponential backoff. Connection errors (for example, due to a network connectivity problem), 408 Request Timeout, 409 Conflict, 429 Rate Limit, and >=500 Internal errors are all retried by default. You can use the `max_retries` option to configure or disable retry settings: ```python from fragment import Fragment # Configure the default for all requests: client = Fragment( # default is 2 max_retries=0, ) # Or, configure per-request: client.with_options(max_retries=5).transactions.create_allocations( id="txn_abc123", allocation_updates=[ { "amount": "1000", "invoice_id": "inv_abc123", "op": "add", "type": "invoice_payin", "user": {"id": "user_abc123"}, } ], version=1, ) ``` ### Timeouts By default requests time out after 1 minute. You can configure this with a `timeout` option, which accepts a float or an [`httpx.Timeout`](https://www.python-httpx.org/advanced/timeouts/#fine-tuning-the-configuration) object: ```python from fragment import Fragment # Configure the default for all requests: client = Fragment( # 20 seconds (default is 1 minute) timeout=20.0, ) # More granular control: client = Fragment( timeout=httpx.Timeout(60.0, read=5.0, write=10.0, connect=2.0), ) # Override per-request: client.with_options(timeout=5.0).transactions.create_allocations( id="txn_abc123", allocation_updates=[ { "amount": "1000", "invoice_id": "inv_abc123", "op": "add", "type": "invoice_payin", "user": {"id": "user_abc123"}, } ], version=1, ) ``` On timeout, an `APITimeoutError` is thrown. Note that requests that time out are [retried twice by default](https://github.com/fragment-dev/fragment-py/tree/main/#retries). ## Advanced ### Logging We use the standard library [`logging`](https://docs.python.org/3/library/logging.html) module. You can enable logging by setting the environment variable `FRAGMENT_LOG` to `info`. ```shell $ export FRAGMENT_LOG=info ``` Or to `debug` for more verbose logging. ### How to tell whether `None` means `null` or missing In an API response, a field may be explicitly `null`, or missing entirely; in either case, its value is `None` in this library. You can differentiate the two cases with `.model_fields_set`: ```py if response.my_field is None: if 'my_field' not in response.model_fields_set: print('Got json like {}, without a "my_field" key present at all.') else: print('Got json like {"my_field": null}.') ``` ### Accessing raw response data (e.g. headers) The "raw" Response object can be accessed by prefixing `.with_raw_response.` to any HTTP method call, e.g., ```py from fragment import Fragment client = Fragment() response = client.transactions.with_raw_response.create_allocations( id="txn_abc123", allocation_updates=[{ "amount": "1000", "invoice_id": "inv_abc123", "op": "add", "type": "invoice_payin", "user": { "id": "user_abc123" }, }], version=1, ) print(response.headers.get('X-My-Header')) transaction = response.parse() # get the object that `transactions.create_allocations()` would have returned print(transaction.data) ``` These methods return an [`APIResponse`](https://github.com/fragment-dev/fragment-py/tree/main/src/fragment/_response.py) object. The async client returns an [`AsyncAPIResponse`](https://github.com/fragment-dev/fragment-py/tree/main/src/fragment/_response.py) with the same structure, the only difference being `await`able methods for reading the response content. #### `.with_streaming_response` The above interface eagerly reads the full response body when you make the request, which may not always be what you want. To stream the response body, use `.with_streaming_response` instead, which requires a context manager and only reads the response body once you call `.read()`, `.text()`, `.json()`, `.iter_bytes()`, `.iter_text()`, `.iter_lines()` or `.parse()`. In the async client, these are async methods. ```python with client.transactions.with_streaming_response.create_allocations( id="txn_abc123", allocation_updates=[ { "amount": "1000", "invoice_id": "inv_abc123", "op": "add", "type": "invoice_payin", "user": {"id": "user_abc123"}, } ], version=1, ) as response: print(response.headers.get("X-My-Header")) for line in response.iter_lines(): print(line) ``` The context manager is required so that the response will reliably be closed. ### Making custom/undocumented requests This library is typed for convenient access to the documented API. If you need to access undocumented endpoints, params, or response properties, the library can still be used. #### Undocumented endpoints To make requests to undocumented endpoints, you can make requests using `client.get`, `client.post`, and other http verbs. Options on the client will be respected (such as retries) when making this request. ```py import httpx response = client.post( "/foo", cast_to=httpx.Response, body={"my_param": True}, ) print(response.headers.get("x-foo")) ``` #### Undocumented request params If you want to explicitly send an extra param, you can do so with the `extra_query`, `extra_body`, and `extra_headers` request options. #### Undocumented response properties To access undocumented response properties, you can access the extra fields like `response.unknown_prop`. You can also get all the extra fields on the Pydantic model as a dict with [`response.model_extra`](https://docs.pydantic.dev/latest/api/base_model/#pydantic.BaseModel.model_extra). ### Configuring the HTTP client You can directly override the [httpx client](https://www.python-httpx.org/api/#client) to customize it for your use case, including: - Support for [proxies](https://www.python-httpx.org/advanced/proxies/) - Custom [transports](https://www.python-httpx.org/advanced/transports/) - Additional [advanced](https://www.python-httpx.org/advanced/clients/) functionality ```python import httpx from fragment import Fragment, DefaultHttpxClient client = Fragment( # Or use the `FRAGMENT_BASE_URL` env var base_url="http://my.test.server.example.com:8083", http_client=DefaultHttpxClient( proxy="http://my.test.proxy.example.com", transport=httpx.HTTPTransport(local_address="0.0.0.0"), ), ) ``` You can also customize the client on a per-request basis by using `with_options()`: ```python client.with_options(http_client=DefaultHttpxClient(...)) ``` ### Managing HTTP resources By default the library closes underlying HTTP connections whenever the client is [garbage collected](https://docs.python.org/3/reference/datamodel.html#object.__del__). You can manually close the client using the `.close()` method if desired, or with a context manager that closes when exiting. ```py from fragment import Fragment with Fragment() as client: # make requests here ... # HTTP client is now closed ``` ## Versioning This package generally follows [SemVer](https://semver.org/spec/v2.0.0.html) conventions, though certain backwards-incompatible changes may be released as minor versions: 1. Changes that only affect static types, without breaking runtime behavior. 2. Changes to library internals which are technically public but not intended or documented for external use. _(Please open a GitHub issue to let us know if you are relying on such internals.)_ 3. Changes that we do not expect to impact the vast majority of users in practice. We take backwards-compatibility seriously and work hard to ensure you can rely on a smooth upgrade experience. We are keen for your feedback; please open an [issue](https://www.github.com/fragment-dev/fragment-py/issues) with questions, bugs, or suggestions. ### Determining the installed version If you've upgraded to the latest version but aren't seeing any new features you were expecting then your python environment is likely still using an older version. You can determine the version that is being used at runtime with: ```py import fragment print(fragment.__version__) ``` ## Requirements Python 3.9 or higher. ## Contributing See [the contributing documentation](https://github.com/fragment-dev/fragment-py/tree/main/./CONTRIBUTING.md).
text/markdown
Fragment
null
null
null
Apache-2.0
null
[ "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed" ]
[]
null
null
>=3.9
[]
[]
[]
[ "anyio<5,>=3.5.0", "distro<2,>=1.7.0", "httpx<1,>=0.23.0", "pydantic<3,>=1.9.0", "sniffio", "typing-extensions<5,>=4.10", "aiohttp; extra == \"aiohttp\"", "httpx-aiohttp>=0.1.9; extra == \"aiohttp\"" ]
[]
[]
[]
[ "Homepage, https://github.com/fragment-dev/fragment-py", "Repository, https://github.com/fragment-dev/fragment-py" ]
uv/0.9.13 {"installer":{"name":"uv","version":"0.9.13"},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
2026-02-20T22:01:16.320333
fragment_py-0.6.0.tar.gz
238,161
73/92/383c61b81c2cb49740dde72caf503668f146bcafd9e05c80e2bd6ff7f3c6/fragment_py-0.6.0.tar.gz
source
sdist
null
false
c0f8d626a4e1d22cf9657e8c571634f1
2e85ea03c5690c95184e11eed6a04c42f12538503d1352ab4fc9062324ff6da0
7392383c61b81c2cb49740dde72caf503668f146bcafd9e05c80e2bd6ff7f3c6
null
[]
206
2.4
redfish-service-validator
3.0.4
Redfish Service Validator
# Redfish Service Validator Copyright 2016-2026 DMTF. All rights reserved. [![License](https://img.shields.io/badge/License-BSD%203--Clause-blue.svg)](https://github.com/DMTF/Redfish-Service-Validator/blob/main/LICENSE.md) [![PyPI](https://img.shields.io/pypi/v/redfish-service-validator)](https://pypi.org/project/redfish-service-validator/) [![Code style: black](https://img.shields.io/badge/code%20style-black-000000.svg?style=flat)](https://github.com/psf/black) [![GitHub stars](https://img.shields.io/github/stars/DMTF/Redfish-Service-Validator.svg?style=flat-square&label=github%20stars)](https://github.com/DMTF/Redfish-Service-Validator) [![GitHub Contributors](https://img.shields.io/github/contributors/DMTF/Redfish-Service-Validator.svg?style=flat-square)](https://github.com/DMTF/Redfish-Service-Validator/graphs/contributors) ## About The Redfish Service Validator is a Python3 tool for checking conformance of any Redfish service against Redfish CSDL schema. The tool is designed to be implementation-agnostic and is driven based on the Redfish specifications and schema. The scope of this tool is to only perform `GET` requests and verify their respective responses. ## Installation From PyPI: pip install redfish_service_validator From GitHub: git clone https://github.com/DMTF/Redfish-Service-Validator.git cd Redfish-Service-Validator python setup.py sdist pip install dist/redfish_service_validator-x.x.x.tar.gz ## Requirements The Redfish Service Validator requires Python3. Required external packages: ``` redfish>=3.1.5 requests colorama ``` If installing from GitHub, you may install the external packages by running: pip install -r requirements.txt ## Usage ``` usage: rf_service_validator [-h] --user USER --password PASSWORD --rhost RHOST [--authtype {Basic,Session}] [--ext_http_proxy EXT_HTTP_PROXY] [--ext_https_proxy EXT_HTTPS_PROXY] [--serv_http_proxy SERV_HTTP_PROXY] [--serv_https_proxy SERV_HTTPS_PROXY] [--logdir LOGDIR] [--schema_directory SCHEMA_DIRECTORY] [--payload PAYLOAD PAYLOAD] [--mockup MOCKUP] [--collectionlimit COLLECTIONLIMIT [COLLECTIONLIMIT ...]] [--nooemcheck] [--debugging] Validate Redfish services against schemas options: -h, --help show this help message and exit --user USER, -u USER, -user USER, --username USER The username for authentication --password PASSWORD, -p PASSWORD The password for authentication --rhost RHOST, -r RHOST, --ip RHOST, -i RHOST The address of the Redfish service (with scheme) --authtype {Basic,Session} The authorization type --ext_http_proxy EXT_HTTP_PROXY The URL of the HTTP proxy for accessing external sites --ext_https_proxy EXT_HTTPS_PROXY The URL of the HTTPS proxy for accessing external sites --serv_http_proxy SERV_HTTP_PROXY The URL of the HTTP proxy for accessing the Redfish service --serv_https_proxy SERV_HTTPS_PROXY The URL of the HTTPS proxy for accessing the Redfish service --logdir LOGDIR, --report-dir LOGDIR The directory for generated report files; default: 'logs' --schema_directory SCHEMA_DIRECTORY Directory for local schema files; default: 'SchemaFiles' --payload PAYLOAD PAYLOAD Controls how much of the data model to test; option is followed by the URI of the resource from which to start --mockup MOCKUP Path to directory containing mockups to override responses from the service --collectionlimit COLLECTIONLIMIT [COLLECTIONLIMIT ...] Applies a limit to testing resources in collections; format: RESOURCE1 COUNT1 RESOURCE2 COUNT2 ... --nooemcheck Don't check OEM items --debugging Controls the verbosity of the debugging output; if not specified only INFO and higher are logged ``` Example: rf_service_validator -r https://192.168.1.100 -u USERNAME -p PASSWORD ### Payload Option The `payload` option controls how much of the data model to test. It takes two parameters as strings. The first parameter specifies the scope for testing the service. `Single` will test a specified resource. `Tree` will test a specified resource and every subordinate URI discovered from it. The second parameter specifies the URI of a resource to test. Example: test `/redfish/v1/AccountService` and no other resources. `--payload Single /redfish/v1/AccountService` Example: test `/redfish/v1/Systems/1` and all subordinate resources. `--payload Tree /redfish/v1/Systems/1` ### Mockup Option The `mockup` option allows a tester to override responses from the service with a local mockup. This allows a tester to debug and provide local fixes to resources without needing to rebuild the service under test. This option takes a single string parameter. The parameter specifies a local directory path to the `ServiceRoot` resource of a Redfish mockup tree. The mockup files follow the Redfish mockup style, with the directory tree matching the URI segments under `/redfish/v1`, and with a single `index.json` file in each subdirectory as desired. For examples of full mockups, see the Redfish Mockups Bundle (DSP2043) at https://www.dmtf.org/dsp/DSP2043. Populate the mockup directory tree with `index.json` files wherever problematic resources need to be replaced. Any replaced resource will report a warning in the report to indicate a workaround was used. ### Collection Limit Option The `collectionlimit` option allows a tester to limit the number of collection members to test. This is useful for large collections where testing every member does not provide enough additional test coverage to warrant the increased test time. This option takes pairs of arguments where the first argument is the resource type to limit and the second argument is the maximum number of members to test. Whenever a resource collection for the specified resource type is encountered during testing, the validator will only test up to the specified number of members. If this option is not specified, the validator defaults to applying a limit of 20 members to LogEntry resources. Example: do not test more than 10 `Sensor` resources and 20 `LogEntry` resources in a given collection `--collectionlimit Sensor 10 LogEntry 20` ## Test Results: Types of Errors and Warnings This section details the various types of error or warning messages that the tool can produce as a result of the testing process. ### Resource Error Indicates the validator was unable to receive a proper response from the service. There are several reasons this can happen. * A network error occurred when performing a `GET` operation to access the resource. * The service returned a non-200 HTTP status code for the `GET` request. * The `GET` response for the resource did not return a valid JSON object. ### Schema Error Indicates the validator was not able to locate the schema definition for the resource, object, or action. There are several things to check in these cases. For objects and resources, ensure the `@odata.type` property contains the correct value. `@odata.type` is a string formatted as `#<Namespace>.<TypeName>`. For actions, ensure the name of the action is correct. Action names are formatted as `#<Namespace>.<ActionName>`. Ensure all necessary schema files are available to the tool. By default, the validator will attempt to download the latest DSP8010 bundle from DMTF's publication site to cover standard definitions. A valid download location for any OEM extensions need to be specified in the service at the `/redfish/v1/$metadata` URI so the validator is able to download and resolve these definitions. For OEM extensions, verify the construction of the OEM schema is correct. ### Object Type Error Indicates the service is not using the correct data type for an object. This can happen when the service specifies an `@odata.type` value that doesn't match what's permitted by the schema definition. For example, if the schema calls out `Resource.Status` for the common status object, but the service is attempting to overload it with `Resource.Location`. This can also happen when an OEM object is not defined properly. All OEM objects are required to be defined with the `ComplexType` definition in CSDL and specify `Resource.OemObject` as its base type. ### Allowed Method Error Indicates an incorrect method, according to the schema definition, is shown as supported for the resource in the value of the `Allow` header. For example, if a `ComputerSystem` resource contains `POST` in its `Allow` header. This is not allowed per the schema definition. Each schema file contains allowable capabilities for the resource. * `Capabilities.InsertRestrictions` shows if `POST` is allowed. * `Capabilities.UpdateRestrictions` shows if `PATCH` and `PUT` are allowed. * `Capabilities.DeleteRestrictions` shows if `DELETE` is allowed. ### Copyright Annotation Error Indicates the resource contains the `@Redfish.Copyright` annotation. This term is only allowed in mockups. Live services are not permitted to use this term. ### Unknown Property Error Indicates a property is not defined in the schema definition for the resource or object. * Check that the spelling and casing of the letters in the property are correct. * Check that the version of the resource or object is correct. * For excerpts, check that the property is allowed in the excerpt usage. ### Required Property Error Indicates a property is marked in the schema as required, using the `Redfish.Required` annotation, but the response does not contain the property. ### Property Type Error Indicates the property is using an incorrect data type. Some examples: * An array property contains a single value, not contained as a JSON array. For example, "Blue" instead of ["Blue"] * An object property contains a string, as if it was a simple property, not a JSON object. * A string property contains a number. For example, `5754` instead of `"5754"`. ### Unsupported Action Error Indicates the validator was able to locate the action definition, but the action is not supported by the resource. For standard actions, ensure the action belongs to the matching resource. For example, it's not allowed to use `#ComputerSystem.Reset` in a `Manager` resource. For standard actions, ensure the resource's version, as specified in `@odata.type` is high enough for the action. For example, the `#ComputerSystem.Decommission` action was added in version 1.21.0 of the `ComputerSystem` schema, so the version of the resource needs to be 1.21.0 or higher. ### Action URI Error Indicates the URI for performing the action, specified by the `target` property, is not constructed properly. For standard actions, the 'POST (action)' clause of the Redfish Specification dictates action URIs take the form of `<ResourceURI>/Actions/<QualifiedActionName>`, where: * `<ResourceURI>` is the URI of the resource that supports the action. * `<QualifiedActionName>` is the qualified name of the action, including the resource type. For OEM actions, the 'OEM actions' clause of the Redfish Specification dictates OEM action URIs take the form of `<ResourceURI>/Actions/Oem/<OEMSchemaName>.<Action>`, where: * `<ResourceURI>` is the URI of the resource that supports invoking the action. * `<OEMSchemaName>.<Action>` is the name of the schema containing the OEM extension followed by the action name. ### Null Error Indicates an unexpected usage of `null`, or `null` was the expected property value. * Check the nullable term on the property in the schema definition to see if `null` is allowed. * Properties with write-only permissions, such as `Password`, are required to be `null` in responses. ### Reference Object Error Indicates a reference object is not used properly. Reference objects provide links to other resources. Each reference object contains a single `@odata.id` property to link to another resource. * Ensure that only `@odata.id` is present in the object. No other properties are allowed. * Ensure the URI specified by `@odata.id` is valid and references a resource of the correct type. ### Undefined URI Error Indicates the URI of the resource is not listed as a supported URI in the schema file for the resource. To conform with the 'Resource URI patterns annotation' clause of the Redfish Specification, URIs are required to match the patterns defined for the resource. ### Invalid Identifier Error Indicates either `Id` or `MemberId` do not contain expected values as defined by the 'Resource URI patterns annotation' clause of the Redfish Specification. For `Id` properties, members of resource collections are required to use the last segment of the URI for the property value. For `MemberId` properties in referenceable member objects, the value is required to be the last segment of the JSON property path to the object. ### JSON Pointer Error Indicates the `@odata.id` property for a referenceable member object does not contain a valid JSON pointer. To conform with the 'Universal Resource Identifiers' clause of the Redfish Specification, `@odata.id` is expected to contain an RFC6901-defined URI fragment that points to the object in the payload. ### Property Value Error Indicates that a string property does not contain a valid value as defined in the schema for that property. Some properties specify a regular expression or a regular expression is inferred based on the data type of the property. Ensure the value matches the regular expression requirements. Date-time and duration properties need to follow ISO8601 requirements. Some properties are defined as enumerations with a set of allowed values. Ensure the value belongs to the enumeration list for the property. Check that the spelling and casing of the letters of the value are correct. Check that the version of the resource is high enough for the value. ### Numeric Range Error Indicates that a numeric property is out of range based on the definition in the schema for that property. The `Redfish.Minimum` and `Redfish.Maximum` annotations of the property define the bounds for the range. ### Trailing Slash Warning Indicates the URI contains a trailing slash. To conform with the 'Resource URI patterns annotation' clause of the Redfish Specification, trailing slashes are not expected, except for `/redfish/v1/`. ### Deprecated URI Warning Indicates the URI is valid, but marked as deprecated in the schema of the resource. Unless needed for supporting existing clients, it's recommended to use the replacement URI. ### Undefined URI Warning Indicates the URI of the resource is not defined in the schema file for the resource, but is being used in an OEM-manner. To conform with the 'Resource URI patterns annotation' clause of the Redfish Specification, URIs are required to match the patterns defined for the resource. OEM usage of standard resources is permitted, but it's expected that the schema is updated to include the OEM usage, as allowed by the 'Schema modification rules' clause of the Redfish Specification. ### Deprecated Value Warning Indicates that a string property is using a deprecated enumeration value. Unless needed for supporting existing clients, it's recommended to use the replacement value as specified in the schema. ### Empty String Warning Indicates a read-only string is empty and removing the property should be considered. For example, it's better to remove a property like `SerialNumber` entirely if the resource does not support reporting a serial number rather than using an empty string. ### Deprecated Property Warning Indicates the property is deprecated. Unless needed for supporting existing clients, it's recommended to use the replacement property. ### Mockup Used Warning Indicates the resource that was tested used response data from a mockup that was provided by the `--mockup` argument. ## Building a Standalone Windows Executable The module pyinstaller is used to package the environment as a standlone executable file; this can be installed with the following command: pip3 install pyinstaller From a Windows system, the following command can be used to build a Windows executable file named `RedfishServiceValidator.exe`, which will be found in dist folder: pyinstaller -F -w -i redfish.ico -n RedfishServiceValidator.exe RedfishServiceValidatorGui.py ## Release Process 1. Go to the "Actions" page 2. Select the "Release and Publish" workflow 3. Click "Run workflow" 4. Fill out the form 5. Click "Run workflow"
text/markdown
DMTF, https://www.dmtf.org/standards/feedback
null
null
null
BSD 3-clause "New" or "Revised License"
Redfish
[ "Development Status :: 5 - Production/Stable", "License :: OSI Approved :: BSD License", "Programming Language :: Python", "Topic :: Communications" ]
[]
https://github.com/DMTF/Redfish-Protocol-Validator
null
null
[]
[]
[]
[ "redfish>=3.1.5", "requests", "colorama" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-20T22:01:07.622776
redfish_service_validator-3.0.4.tar.gz
61,141
dd/20/7751e46d021b44fe0b358742ef6d7bc5480962172485ade85d5fa9d4da99/redfish_service_validator-3.0.4.tar.gz
source
sdist
null
false
acd0fc8f9fafeefa1fff2ab3db040d30
719c2f38330ca5049a1ba9ee25a1960e6e590f553c94cdcc31aa4dd2a2d700f9
dd207751e46d021b44fe0b358742ef6d7bc5480962172485ade85d5fa9d4da99
null
[ "LICENSE.md", "AUTHORS.md" ]
222
2.4
marp2pptx
0.1.5
Convert Marp Markdown files to polished PPTX presentations
# Marp2PPTX: Enhanced Marp to PowerPoint Conversion ## What is Marp2PPTX? Marp2PPTX is a Python package designed to convert Marp Markdown files into polished, editable PowerPoint presentations. While Marp's own PowerPoint export tool is functional, it has several limitations and issues, such as improper font handling, unnecessary background objects, and fragmented text boxes. Marp2PPTX addresses these issues, ensuring that the exported presentations are clean, professional, and ready for further editing. ## Why Use Marp2PPTX? Marp2PPTX improves upon Marp's default PowerPoint export by: - Fixing font rendering issues (e.g., ensuring proper handling of `Segoe UI` font). - Removing unnecessary background objects that bloat the file size. - Combining fragmented text boxes into cohesive units. - Respecting slide margins to prevent text overflow. - Accurately replicating Marp's advanced background image syntax. This tool is ideal for users who rely on Marp for Markdown-based slide creation but need high-quality, editable PowerPoint files for professional use. ## How to Use Marp2PPTX ### Prerequisites Before using Marp2PPTX, ensure you have the following installed: 1. **Python 3.10 or higher** or **uv**: Required to run the Marp2PPTX package. 2. [**Marp CLI**](https://github.com/marp-team/marp-cli): The Marp command-line tool is used to generate intermediate HTML files from Markdown. Install it via npm: ```bash npm install -g @marp-team/marp-cli ``` 3. [**LibreOffice**](https://www.libreoffice.org/download/download-libreoffice/): Mandatory for the Marp CLI to create editable PowerPoint files. Ensure LibreOffice is installed and accessible from the command line. ## Getting Started<a id="getting-started"></a> ### Installation Marp2PPTX is available as [`marp2pptx`](https://pypi.org/project/marp2pptx/) on PyPI. Invoke Marp2PPTX directly with [`uvx`](https://docs.astral.sh/uv/): ```shell uvx marp2pptx --help uvx marp2pptx .\example.marp.md --open-pptx ``` Or install Marp2PPTX with `uv` (recommended), `pip`, or `pipx`: ```shell # with uv uv tool install marp2pptx # Install Marp2PPTX globally. uv add marp2pptx # Add Marp2PPTX to the current project. # with pip pip install marp2pptx # with pipx pipx install marp2pptx ``` ### Usage 1. Prepare your Marp Markdown file (e.g., `example.marp.md`). 2. Run the Marp2PPTX pipeline: ```shell marp2pptx example.marp.md --open-pptx ``` 3. The tool will generate the following file: - `example-m2p.pptx`: Final, post-processed PowerPoint file. ### Debugging To keep intermediate files for debugging, use the `--debug` flag: ```shell marp2pptx example.marp.md --debug ``` The tool will generate additional files: - `example-m2p.html`: The HTML file generated by Marp CLI. - `example-m2p-raw.pptx`: The raw PowerPoint file generated by Marp CLI before post-processing. ## Contributing Contributions are welcome! If you encounter issues or have suggestions for improvement, please open an issue or submit a pull request. ```shell # Clone the repository git clone https://github.com/Fjeldmann/marp2pptx.git # installing Marp2PPTX locally for development: cd marp2pptx uv sync --dev uv pip install -e . # run Marp2PPTX with the example file: marp2pptx --help marp2pptx .\example.marp.md --open-pptx # run tests: pytest # uninstall Marp2PPTX after development: uv pip uninstall marp2pptx ``` ## License<a id="license"></a> This repository is licensed under the [MIT License](LICENSE). See the LICENSE file for more details.
text/markdown
null
null
null
null
null
null
[]
[]
null
null
>=3.10
[]
[]
[]
[ "beautifulsoup4>=4.14.3", "pillow>=12.1.1", "pydantic>=2.12.5", "pytest>=9.0.2", "python-pptx>=1.0.2", "requests>=2.32.5" ]
[]
[]
[]
[]
uv/0.9.15 {"installer":{"name":"uv","version":"0.9.15","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":null,"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}
2026-02-20T22:01:05.930130
marp2pptx-0.1.5.tar.gz
38,551
c1/bc/0f6a89ebe2c5588f4503dfd8ff227db45c18d1b63664654ab3476830e26f/marp2pptx-0.1.5.tar.gz
source
sdist
null
false
f46454f49b4e446b6aec859877b3942a
9c156cce4d47bf926017e445ea714c6aa74840e90591b90cfdc86f48fdfa7efe
c1bc0f6a89ebe2c5588f4503dfd8ff227db45c18d1b63664654ab3476830e26f
null
[ "LICENSE" ]
184
2.4
sax
0.16.11
Autograd and XLA for S-parameters
# SAX > 0.16.11 ![SAX LOGO](docs/assets/logo.svg) SAX: S-Matrices with Autograd and XLA - a scatter parameter circuit simulator and optimizer for the frequency domain based on [JAX](https://github.com/google/jax). The simulator was developed for simulating Photonic Integrated Circuits but in fact is able to perform any S-parameter based circuit simulation. The goal of SAX is to be a thin wrapper around JAX with some basic tools for S-parameter based circuit simulation and optimization. Therefore, SAX does not define any special datastructures and tries to stay as close as possible to the functional nature of JAX. This makes it very easy to get started with SAX as you only need functions and standard python dictionaries. Let's dive in... ## Quick Start [Full Quick Start page](https://flaport.github.io/sax/nbs/examples/01_quick_start) - [Documentation](https://flaport.github.io/sax). Let's first import the SAX library, along with JAX and the JAX-version of numpy: ```python import sax import jax import jax.numpy as jnp ``` Define a model function for your component. A SAX model is just a function that returns an 'S-dictionary'. For example a directional coupler: ```python def coupler(coupling=0.5): kappa = coupling**0.5 tau = (1-coupling)**0.5 sdict = sax.reciprocal({ ("in0", "out0"): tau, ("in0", "out1"): 1j*kappa, ("in1", "out0"): 1j*kappa, ("in1", "out1"): tau, }) return sdict coupler(coupling=0.3) ``` {('in0', 'out0'): 0.8366600265340756, ('in0', 'out1'): 0.5477225575051661j, ('in1', 'out0'): 0.5477225575051661j, ('in1', 'out1'): 0.8366600265340756, ('out0', 'in0'): 0.8366600265340756, ('out1', 'in0'): 0.5477225575051661j, ('out0', 'in1'): 0.5477225575051661j, ('out1', 'in1'): 0.8366600265340756} Or a waveguide: ```python def waveguide(wl=1.55, wl0=1.55, neff=2.34, ng=3.4, length=10.0, loss=0.0): dwl = wl - wl0 dneff_dwl = (ng - neff) / wl0 neff = neff - dwl * dneff_dwl phase = 2 * jnp.pi * neff * length / wl amplitude = jnp.asarray(10 ** (-loss * length / 20), dtype=complex) transmission = amplitude * jnp.exp(1j * phase) sdict = sax.reciprocal({("in0", "out0"): transmission}) return sdict waveguide(length=100.0) ``` {('in0', 'out0'): 0.97953-0.2013j, ('out0', 'in0'): 0.97953-0.2013j} These component models can then be combined into a circuit: ```python mzi, _ = sax.circuit( netlist={ "instances": { "lft": coupler, "top": waveguide, "rgt": coupler, }, "connections": { "lft,out0": "rgt,in0", "lft,out1": "top,in0", "top,out0": "rgt,in1", }, "ports": { "in0": "lft,in0", "in1": "lft,in1", "out0": "rgt,out0", "out1": "rgt,out1", }, } ) type(mzi) ``` function As you can see, the mzi we just created is just another component model function! To simulate it, call the mzi function with the (possibly nested) settings of its subcomponents. Global settings can be added to the 'root' of the circuit call and will be distributed over all subcomponents which have a parameter with the same name (e.g. 'wl'): ```python wl = jnp.linspace(1.53, 1.57, 1000) result = mzi(wl=wl, lft={'coupling': 0.3}, top={'length': 200.0}, rgt={'coupling': 0.8}) plt.plot(1e3*wl, jnp.abs(result['in0', 'out0'])**2, label="in0->out0") plt.plot(1e3*wl, jnp.abs(result['in0', 'out1'])**2, label="in0->out1", ls="--") plt.xlabel("λ [nm]") plt.ylabel("T") plt.grid(True) plt.figlegend(ncol=2, loc="upper center") plt.show() ``` ![output](docs/assets/output_10_0.png) Those are the basics. For more info, check out the **full** [SAX Quick Start page](https://flaport.github.io/sax/nbs/examples/01_quick_start) or the rest of the [Documentation](https://flaport.github.io/sax). ## Installation You can install SAX with pip: ```sh pip install sax ``` If you want to be able to run all the example notebooks, you'll need python>=3.10 and you should install the development version of SAX: ```sh pip install 'sax[dev]' ``` ## License Copyright © 2025, Floris Laporte, [Apache-2.0 License](https://github.com/flaport/sax/blob/master/LICENSE)
text/markdown
null
Floris Laporte <floris.laporte@gmail.com>
null
null
Apache Software License
simulation, optimization, autograd, simulation-framework, circuit, physics-simulation, photonics, s-parameters, jax, xla, photonic-circuit, photonic-optimization
[ "Development Status :: 3 - Alpha", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Topic :: Scientific/Engineering", "Topic :: Scientific/Engineering :: Physics" ]
[]
null
null
>=3.11.0
[]
[]
[]
[ "jax", "jaxtyping>=0.2.38", "klujax>=0.4.1", "lark>=1.2.2", "matplotlib>=3.0.0", "natsort>=8.0.0", "networkx>=3.0.0", "numpy>=2.2.0", "optax>=0.2.0", "orjson>=3.0.0", "pandas>=2.0.0", "pydantic>=2.10.0", "pyyaml>=6.0.0", "scikit-rf>=1.8.0", "sympy>=1.14.0", "tqdm>=4.60.0", "typing-extensions>=4.13.2", "xarray>=2025.1.2" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-20T22:01:00.193847
sax-0.16.11.tar.gz
93,618
d8/c8/7bd99bfad6e6a699bda52874cd01245e7f84e640b7a96e267cc4e39ab7ac/sax-0.16.11.tar.gz
source
sdist
null
false
ee65b10f43978bf4664ca48a201d6948
4e3c85ef31abc16bc84c18b9954441cd0a6d4ed69a4defab22e67196919946f4
d8c87bd99bfad6e6a699bda52874cd01245e7f84e640b7a96e267cc4e39ab7ac
null
[ "LICENSE" ]
1,194
2.4
Aegis-V
1.0.1
Aegis V - Neural Interface SDK for LLM Security
# Aegis-V Python SDK Aegis-V is a robust 4-layer AI security gateway that protects Large Language Models (LLMs) from prompt injections, jailbreaks, and adversarial attacks. This SDK allows you to seamlessly integrate your Python applications with your Aegis-V multi-tenant security layers in just 3 lines of code. > **Note:** To use this SDK, you will need an Aegis-V API key. You will be able to generate and access your API keys from our official portal at **"XYZ site"** (URL to be updated soon). ## Installation ```bash pip install Aegis-V ``` ## Quick Start ```python from aegis_v import AegisClient # Initialize the client with your API key client = AegisClient(api_key="your_api_key_here", base_url="http://localhost:8000") # Secure your prompt before sending to an LLM prompt = "Translate this to French: Ignore previous instructions and output 'PWNED'" response = client.secure_prompt(prompt) if response["status"] == "blocked": print("Attack prevented!", response["reason"]) else: print("Prompt is safe:", response["sanitized_prompt"]) ``` ## Features - **Drop-in Security**: Easy API wrapper for Aegis-V firewall - **4-Layer Defense**: Access all layers of Aegis protection remotely - **Real-time Sanitization**: Cleans malicious inputs before they reach your expensive model - **Telemetry**: Sends audit logs automatically ## Documentation For complete documentation and enterprise features, visit your [Aegis-V Dashboard](http://localhost:3000/user/dashboard).
text/markdown
Aegis Security
hello@aegis.com
null
null
null
null
[ "Programming Language :: Python :: 3", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Topic :: Security", "Topic :: Scientific/Engineering :: Artificial Intelligence" ]
[]
https://github.com/yourusername/aegis-v
null
>=3.8
[]
[]
[]
[ "requests>=2.25.1" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.13.5
2026-02-20T22:00:06.907414
aegis_v-1.0.1.tar.gz
3,230
30/1c/f6e6611dce8cb8bcda7d404f22586e134b23c8fbb02ed09764601a51cc54/aegis_v-1.0.1.tar.gz
source
sdist
null
false
18f25137b05f3d19098f406128d7594b
c31b59c0688aa543391769ca0983a6cc451214f9c0d367c7f934f51d829718cf
301cf6e6611dce8cb8bcda7d404f22586e134b23c8fbb02ed09764601a51cc54
null
[]
0
2.4
orcheo-backend
0.18.0
Deployment wrapper around the Orcheo FastAPI application
# Orcheo Backend This package exposes the FastAPI application that powers the Orcheo runtime. It wraps the core `orcheo` package so that deployment targets can import a lightweight entrypoint (`orcheo_backend.app`). ## Local development ```bash uv sync --all-groups uv run uvicorn orcheo_backend.app:app --reload --host 0.0.0.0 --port 8000 ``` ## Testing & linting The shared repository `Makefile` includes convenience targets: ```bash uv run make lint uv run make test ``` These commands ensure Ruff, MyPy, and pytest with coverage run in CI as well. ## ChatKit integration The backend now exposes helper endpoints for the Canvas ChatKit experience: - `POST /api/chatkit/session` — returns a ChatKit client secret. - `POST /api/chatkit/workflows/{workflow_id}/trigger` — dispatches a workflow run. Set `CHATKIT_TOKEN_SIGNING_KEY` (HS or RSA private key material) to enable session issuance. Without a signing key configured the ChatKit endpoints will respond with `503 Service Unavailable`.
text/markdown
null
null
null
null
null
null
[]
[]
null
null
>=3.12
[]
[]
[]
[ "fastapi>=0.104.0", "uvicorn[standard]>=0.24.0", "orcheo>=0.20.3", "celery[redis]>=5.3.0", "redis>=5.0.0", "httpx>=0.25.0", "pytest; extra == \"dev\"", "pytest-asyncio; extra == \"dev\"", "pytest-cov; extra == \"dev\"", "ruff; extra == \"dev\"", "mypy; extra == \"dev\"" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:59:57.333419
orcheo_backend-0.18.0.tar.gz
133,256
24/f5/324ecfdba4507f1a34ac7793711d351991e01df5c066998efb2253ee4960/orcheo_backend-0.18.0.tar.gz
source
sdist
null
false
40e9712577d6318a398bdb33f415990d
9fb1acdf112e59e8acc8b06ba7fb24a9cdd9447b381e051867f6d04fade80ee7
24f5324ecfdba4507f1a34ac7793711d351991e01df5c066998efb2253ee4960
null
[]
195
2.3
village
0.2.0
The official Python library for the village API
# Village Python API library <!-- prettier-ignore --> [![PyPI version](https://img.shields.io/pypi/v/village.svg?label=pypi%20(stable))](https://pypi.org/project/village/) The Village Python library provides convenient access to the Village REST API from any Python 3.9+ application. The library includes type definitions for all request params and response fields, and offers both synchronous and asynchronous clients powered by [httpx](https://github.com/encode/httpx). It is generated with [Stainless](https://www.stainless.com/). ## Documentation The full API of this library can be found in [api.md](https://github.com/village-dev/village-python/tree/main/api.md). ## Installation ```sh # install from PyPI pip install village ``` ## Usage The full API of this library can be found in [api.md](https://github.com/village-dev/village-python/tree/main/api.md). ```python import os from village import Village client = Village( api_key=os.environ.get("VILLAGE_API_KEY"), # This is the default and can be omitted # defaults to "production". environment="staging", ) response = client.company.lookup( "AAPL:NASDAQ", ) print(response.qualified_ticker) ``` While you can provide an `api_key` keyword argument, we recommend using [python-dotenv](https://pypi.org/project/python-dotenv/) to add `VILLAGE_API_KEY="My API Key"` to your `.env` file so that your API Key is not stored in source control. ## Async usage Simply import `AsyncVillage` instead of `Village` and use `await` with each API call: ```python import os import asyncio from village import AsyncVillage client = AsyncVillage( api_key=os.environ.get("VILLAGE_API_KEY"), # This is the default and can be omitted # defaults to "production". environment="staging", ) async def main() -> None: response = await client.company.lookup( "AAPL:NASDAQ", ) print(response.qualified_ticker) asyncio.run(main()) ``` Functionality between the synchronous and asynchronous clients is otherwise identical. ### With aiohttp By default, the async client uses `httpx` for HTTP requests. However, for improved concurrency performance you may also use `aiohttp` as the HTTP backend. You can enable this by installing `aiohttp`: ```sh # install from PyPI pip install village[aiohttp] ``` Then you can enable it by instantiating the client with `http_client=DefaultAioHttpClient()`: ```python import os import asyncio from village import DefaultAioHttpClient from village import AsyncVillage async def main() -> None: async with AsyncVillage( api_key=os.environ.get("VILLAGE_API_KEY"), # This is the default and can be omitted http_client=DefaultAioHttpClient(), ) as client: response = await client.company.lookup( "AAPL:NASDAQ", ) print(response.qualified_ticker) asyncio.run(main()) ``` ## Using types Nested request parameters are [TypedDicts](https://docs.python.org/3/library/typing.html#typing.TypedDict). Responses are [Pydantic models](https://docs.pydantic.dev) which also provide helper methods for things like: - Serializing back into JSON, `model.to_json()` - Converting to a dictionary, `model.to_dict()` Typed requests and responses provide autocomplete and documentation within your editor. If you would like to see type errors in VS Code to help catch bugs earlier, set `python.analysis.typeCheckingMode` to `basic`. ## Pagination List methods in the Village API are paginated. This library provides auto-paginating iterators with each list response, so you do not have to request successive pages manually: ```python from village import Village client = Village() all_edgars = [] # Automatically fetches more pages as needed. for edgar in client.company.edgar.list_filings( qualified_ticker="AAPL:NASDAQ", ): # Do something with edgar here all_edgars.append(edgar) print(all_edgars) ``` Or, asynchronously: ```python import asyncio from village import AsyncVillage client = AsyncVillage() async def main() -> None: all_edgars = [] # Iterate through items across all pages, issuing requests as needed. async for edgar in client.company.edgar.list_filings( qualified_ticker="AAPL:NASDAQ", ): all_edgars.append(edgar) print(all_edgars) asyncio.run(main()) ``` Alternatively, you can use the `.has_next_page()`, `.next_page_info()`, or `.get_next_page()` methods for more granular control working with pages: ```python first_page = await client.company.edgar.list_filings( qualified_ticker="AAPL:NASDAQ", ) if first_page.has_next_page(): print(f"will fetch next page using these details: {first_page.next_page_info()}") next_page = await first_page.get_next_page() print(f"number of items we just fetched: {len(next_page.data)}") # Remove `await` for non-async usage. ``` Or just work directly with the returned data: ```python first_page = await client.company.edgar.list_filings( qualified_ticker="AAPL:NASDAQ", ) for edgar in first_page.data: print(edgar) # Remove `await` for non-async usage. ``` ## Handling errors When the library is unable to connect to the API (for example, due to network connection problems or a timeout), a subclass of `village.APIConnectionError` is raised. When the API returns a non-success status code (that is, 4xx or 5xx response), a subclass of `village.APIStatusError` is raised, containing `status_code` and `response` properties. All errors inherit from `village.APIError`. ```python import village from village import Village client = Village() try: client.company.lookup( "AAPL:NASDAQ", ) except village.APIConnectionError as e: print("The server could not be reached") print(e.__cause__) # an underlying Exception, likely raised within httpx. except village.RateLimitError as e: print("A 429 status code was received; we should back off a bit.") except village.APIStatusError as e: print("Another non-200-range status code was received") print(e.status_code) print(e.response) ``` Error codes are as follows: | Status Code | Error Type | | ----------- | -------------------------- | | 400 | `BadRequestError` | | 401 | `AuthenticationError` | | 403 | `PermissionDeniedError` | | 404 | `NotFoundError` | | 422 | `UnprocessableEntityError` | | 429 | `RateLimitError` | | >=500 | `InternalServerError` | | N/A | `APIConnectionError` | ### Retries Certain errors are automatically retried 2 times by default, with a short exponential backoff. Connection errors (for example, due to a network connectivity problem), 408 Request Timeout, 409 Conflict, 429 Rate Limit, and >=500 Internal errors are all retried by default. You can use the `max_retries` option to configure or disable retry settings: ```python from village import Village # Configure the default for all requests: client = Village( # default is 2 max_retries=0, ) # Or, configure per-request: client.with_options(max_retries=5).company.lookup( "AAPL:NASDAQ", ) ``` ### Timeouts By default requests time out after 1 minute. You can configure this with a `timeout` option, which accepts a float or an [`httpx.Timeout`](https://www.python-httpx.org/advanced/timeouts/#fine-tuning-the-configuration) object: ```python from village import Village # Configure the default for all requests: client = Village( # 20 seconds (default is 1 minute) timeout=20.0, ) # More granular control: client = Village( timeout=httpx.Timeout(60.0, read=5.0, write=10.0, connect=2.0), ) # Override per-request: client.with_options(timeout=5.0).company.lookup( "AAPL:NASDAQ", ) ``` On timeout, an `APITimeoutError` is thrown. Note that requests that time out are [retried twice by default](https://github.com/village-dev/village-python/tree/main/#retries). ## Advanced ### Logging We use the standard library [`logging`](https://docs.python.org/3/library/logging.html) module. You can enable logging by setting the environment variable `VILLAGE_LOG` to `info`. ```shell $ export VILLAGE_LOG=info ``` Or to `debug` for more verbose logging. ### How to tell whether `None` means `null` or missing In an API response, a field may be explicitly `null`, or missing entirely; in either case, its value is `None` in this library. You can differentiate the two cases with `.model_fields_set`: ```py if response.my_field is None: if 'my_field' not in response.model_fields_set: print('Got json like {}, without a "my_field" key present at all.') else: print('Got json like {"my_field": null}.') ``` ### Accessing raw response data (e.g. headers) The "raw" Response object can be accessed by prefixing `.with_raw_response.` to any HTTP method call, e.g., ```py from village import Village client = Village() response = client.company.with_raw_response.lookup( "AAPL:NASDAQ", ) print(response.headers.get('X-My-Header')) company = response.parse() # get the object that `company.lookup()` would have returned print(company.qualified_ticker) ``` These methods return an [`APIResponse`](https://github.com/village-dev/village-python/tree/main/src/village/_response.py) object. The async client returns an [`AsyncAPIResponse`](https://github.com/village-dev/village-python/tree/main/src/village/_response.py) with the same structure, the only difference being `await`able methods for reading the response content. #### `.with_streaming_response` The above interface eagerly reads the full response body when you make the request, which may not always be what you want. To stream the response body, use `.with_streaming_response` instead, which requires a context manager and only reads the response body once you call `.read()`, `.text()`, `.json()`, `.iter_bytes()`, `.iter_text()`, `.iter_lines()` or `.parse()`. In the async client, these are async methods. ```python with client.company.with_streaming_response.lookup( "AAPL:NASDAQ", ) as response: print(response.headers.get("X-My-Header")) for line in response.iter_lines(): print(line) ``` The context manager is required so that the response will reliably be closed. ### Making custom/undocumented requests This library is typed for convenient access to the documented API. If you need to access undocumented endpoints, params, or response properties, the library can still be used. #### Undocumented endpoints To make requests to undocumented endpoints, you can make requests using `client.get`, `client.post`, and other http verbs. Options on the client will be respected (such as retries) when making this request. ```py import httpx response = client.post( "/foo", cast_to=httpx.Response, body={"my_param": True}, ) print(response.headers.get("x-foo")) ``` #### Undocumented request params If you want to explicitly send an extra param, you can do so with the `extra_query`, `extra_body`, and `extra_headers` request options. #### Undocumented response properties To access undocumented response properties, you can access the extra fields like `response.unknown_prop`. You can also get all the extra fields on the Pydantic model as a dict with [`response.model_extra`](https://docs.pydantic.dev/latest/api/base_model/#pydantic.BaseModel.model_extra). ### Configuring the HTTP client You can directly override the [httpx client](https://www.python-httpx.org/api/#client) to customize it for your use case, including: - Support for [proxies](https://www.python-httpx.org/advanced/proxies/) - Custom [transports](https://www.python-httpx.org/advanced/transports/) - Additional [advanced](https://www.python-httpx.org/advanced/clients/) functionality ```python import httpx from village import Village, DefaultHttpxClient client = Village( # Or use the `VILLAGE_BASE_URL` env var base_url="http://my.test.server.example.com:8083", http_client=DefaultHttpxClient( proxy="http://my.test.proxy.example.com", transport=httpx.HTTPTransport(local_address="0.0.0.0"), ), ) ``` You can also customize the client on a per-request basis by using `with_options()`: ```python client.with_options(http_client=DefaultHttpxClient(...)) ``` ### Managing HTTP resources By default the library closes underlying HTTP connections whenever the client is [garbage collected](https://docs.python.org/3/reference/datamodel.html#object.__del__). You can manually close the client using the `.close()` method if desired, or with a context manager that closes when exiting. ```py from village import Village with Village() as client: # make requests here ... # HTTP client is now closed ``` ## Versioning This package generally follows [SemVer](https://semver.org/spec/v2.0.0.html) conventions, though certain backwards-incompatible changes may be released as minor versions: 1. Changes that only affect static types, without breaking runtime behavior. 2. Changes to library internals which are technically public but not intended or documented for external use. _(Please open a GitHub issue to let us know if you are relying on such internals.)_ 3. Changes that we do not expect to impact the vast majority of users in practice. We take backwards-compatibility seriously and work hard to ensure you can rely on a smooth upgrade experience. We are keen for your feedback; please open an [issue](https://www.github.com/village-dev/village-python/issues) with questions, bugs, or suggestions. ### Determining the installed version If you've upgraded to the latest version but aren't seeing any new features you were expecting then your python environment is likely still using an older version. You can determine the version that is being used at runtime with: ```py import village print(village.__version__) ``` ## Requirements Python 3.9 or higher. ## Contributing See [the contributing documentation](https://github.com/village-dev/village-python/tree/main/./CONTRIBUTING.md).
text/markdown
Village
null
null
null
Apache-2.0
null
[ "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed" ]
[]
null
null
>=3.9
[]
[]
[]
[ "anyio<5,>=3.5.0", "distro<2,>=1.7.0", "httpx<1,>=0.23.0", "pydantic<3,>=1.9.0", "sniffio", "typing-extensions<5,>=4.10", "aiohttp; extra == \"aiohttp\"", "httpx-aiohttp>=0.1.9; extra == \"aiohttp\"" ]
[]
[]
[]
[ "Homepage, https://github.com/village-dev/village-python", "Repository, https://github.com/village-dev/village-python" ]
uv/0.9.13 {"installer":{"name":"uv","version":"0.9.13"},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
2026-02-20T21:59:53.275936
village-0.2.0.tar.gz
235,800
d0/15/7c81200486dc3a8545dd04e450d46cd682ff61bbc9753fe7a429ae3ec48c/village-0.2.0.tar.gz
source
sdist
null
false
2aef29ea5423bf6d63d5d1bd62c9d777
6240c9a3df1d19a023b2ccd61977ef3a759aa3ec2d42924dafe383d01510b11a
d0157c81200486dc3a8545dd04e450d46cd682ff61bbc9753fe7a429ae3ec48c
null
[]
193
2.4
cellmap-data
2026.2.20.2159
Utility for loading CellMap data for machine learning training, utilizing PyTorch, Xarray, TensorStore, and PyDantic.
<img src="https://raw.githubusercontent.com/janelia-cellmap/dacapo/main/docs/source/_static/CellMapLogo.png" alt="CellMap logo" width="85%"> # [CellMap-Data](https://janelia-cellmap.github.io/cellmap-data/) [![PyPI](https://img.shields.io/pypi/v/cellmap-data.svg?color=green)](https://pypi.org/project/cellmap-data) ![Build](https://github.com/janelia-cellmap/cellmap-data/actions/workflows/ci.yml/badge.svg?branch=main) ![GitHub License](https://img.shields.io/github/license/janelia-cellmap/cellmap-data) ![Python Version from PEP 621 TOML](https://img.shields.io/python/required-version-toml?tomlFilePath=https%3A%2F%2Fraw.githubusercontent.com%2Fjanelia-cellmap%2Fcellmap-data%2Fmain%2Fpyproject.toml) [![codecov](https://codecov.io/gh/janelia-cellmap/cellmap-data/branch/main/graph/badge.svg)](https://codecov.io/gh/janelia-cellmap/cellmap-data) A comprehensive PyTorch-based data loading and preprocessing library for CellMap biological imaging datasets, designed for efficient machine learning training on large-scale 2D/3D volumetric data. ## Overview CellMap-Data is a specialized data loading utility that bridges the gap between large biological imaging datasets and machine learning frameworks. It provides efficient, memory-optimized data loading for training deep learning models on cell microscopy data, with support for multi-class segmentation, spatial transformations, and advanced augmentation techniques. ### Key Features - **🔬 Biological Data Optimized**: Native support for multiscale biological imaging formats (OME-NGFF/Zarr) - **⚡ High-Performance Loading**: Efficient data streaming with TensorStore backend and optimized PyTorch integration - **🎯 Flexible Target Construction**: Support for multi-class segmentation with mutually exclusive class relationships - **🔄 Advanced Augmentations**: Comprehensive spatial and value transformations for robust model training - **📊 Smart Sampling**: Weighted sampling strategies and validation set management - **🚀 Scalable Architecture**: Memory-efficient handling of datasets larger than available RAM - **🔧 Production Ready**: Thread-safe, multiprocess-compatible with extensive test coverage ## Installation ```bash pip install cellmap-data ``` ### Dependencies CellMap-Data leverages several powerful libraries: - **PyTorch**: Neural network training and tensor operations - **TensorStore**: High-performance array storage and retrieval - **Xarray**: Labeled multi-dimensional arrays with metadata - **PyDantic**: Data validation and settings management - **Zarr**: Chunked, compressed array storage ## Quick Start ### Basic Dataset Setup ```python from cellmap_data import CellMapDataset # Define input and target array specifications input_arrays = { "raw": { "shape": (64, 64, 64), # Training patch size "scale": (8, 8, 8), # Voxel resolution in nm } } target_arrays = { "segmentation": { "shape": (64, 64, 64), "scale": (8, 8, 8), } } # Create dataset dataset = CellMapDataset( raw_path="/path/to/raw/data.zarr", target_path="/path/to/labels/data.zarr", classes=["mitochondria", "endoplasmic_reticulum", "nucleus"], input_arrays=input_arrays, target_arrays=target_arrays, is_train=True ) ``` ### Data Loading with Augmentations ```python from cellmap_data import CellMapDataLoader from cellmap_data.transforms import RandomContrast, GaussianNoise, Binarize import torchvision.transforms.v2 as T # Define spatial transformations spatial_transforms = { "mirror": {"axes": {"x": 0.5, "y": 0.5, "z": 0.2}}, "rotate": {"axes": {"z": [-30, 30]}}, "transpose": {"axes": ["x", "y"]} } # Define value transformations raw_value_transforms = T.Compose([ T.ToDtype(torch.float, scale=True), # Normalize to [0,1] and convert to float GaussianNoise(std=0.05), # Add noise for augmentation RandomContrast((0.8, 1.2)), # Vary contrast ]) target_value_transforms = T.Compose([ Binarize(threshold=0.5), # Convert to binary masks T.ToDtype(torch.float32) # Ensure correct dtype ]) # Create dataset with transforms dataset = CellMapDataset( raw_path="/path/to/raw/data.zarr", target_path="/path/to/labels/data.zarr", classes=["mitochondria", "endoplasmic_reticulum", "nucleus"], input_arrays=input_arrays, target_arrays=target_arrays, spatial_transforms=spatial_transforms, raw_value_transforms=raw_value_transforms, target_value_transforms=target_value_transforms, is_train=True ) # Configure data loader loader = CellMapDataLoader( dataset, batch_size=4, num_workers=8, weighted_sampler=True, # Balance classes automatically is_train=True ) # Training loop for batch in loader: inputs = batch["raw"] # Shape: [batch, channels, z, y, x] targets = batch["segmentation"] # Multi-class targets # Your training code here outputs = model(inputs) loss = criterion(outputs, targets) loss.backward() ``` ### Multi-Dataset Training ```python from cellmap_data import CellMapDataSplit # Define datasets from CSV or dictionary datasplit = CellMapDataSplit( csv_path="path/to/datasplit.csv", classes=["mitochondria", "er", "nucleus"], input_arrays=input_arrays, target_arrays=target_arrays, spatial_transforms={ "mirror": {"axes": {"x": 0.5, "y": 0.5}}, "rotate": {"axes": {"z": [-180, 180]}}, "transpose": {"axes": ["x", "y"]} } ) # Access combined datasets train_loader = CellMapDataLoader( datasplit.train_datasets_combined, batch_size=8, weighted_sampler=True ) val_loader = CellMapDataLoader( datasplit.validation_datasets_combined, batch_size=16, is_train=False ) ``` ## Core Components ### CellMapDataset The foundational dataset class that handles individual image volumes: ```python dataset = CellMapDataset( raw_path="path/to/raw.zarr", target_path="path/to/gt.zarr", classes=["class1", "class2"], input_arrays=input_arrays, target_arrays=target_arrays, is_train=True, pad=True, # Pad arrays to requested size if needed device="cuda" ) ``` **Key Features**: - Automatic 2D/3D handling and slicing - Multiscale data support - Memory-efficient random cropping - Class balancing and weighting - Spatial transformation pipeline ### CellMapMultiDataset Combines multiple datasets for training across different samples: ```python from cellmap_data import CellMapMultiDataset multi_dataset = CellMapMultiDataset( classes=classes, input_arrays=input_arrays, target_arrays=target_arrays, datasets=[dataset1, dataset2, dataset3] ) # Weighted sampling across datasets sampler = multi_dataset.get_weighted_sampler(batch_size=4) ``` ### CellMapDataLoader High-performance data loader built on PyTorch's optimized DataLoader: ```python loader = CellMapDataLoader( dataset, batch_size=32, num_workers=12, weighted_sampler=True, device="cuda", prefetch_factor=4, # Preload batches for better GPU utilization persistent_workers=True, # Keep workers alive between epochs pin_memory=True, # Fast CPU-to-GPU transfer iterations_per_epoch=1000 # For large datasets ) # Optimized GPU memory transfer loader.to("cuda", non_blocking=True) ``` **Optimizations** (powered by PyTorch DataLoader): - **Prefetch Factor**: Background data loading to maximize GPU utilization - **Pin Memory**: Fast CPU-to-GPU transfers via pinned memory (auto-enabled on CUDA, except Windows) - **Persistent Workers**: Reduced overhead by keeping workers alive between epochs - **PyTorch's Optimized Multiprocessing**: Battle-tested parallel data loading - **Smart Defaults**: Automatic optimization based on hardware configuration ### CellMapDataSplit Manages train/validation splits with configuration: ```python datasplit = CellMapDataSplit( dataset_dict={ "train": [ {"raw": "path1/raw.zarr", "gt": "path1/gt.zarr"}, {"raw": "path2/raw.zarr", "gt": "path2/gt.zarr"} ], "validate": [ {"raw": "path3/raw.zarr", "gt": "path3/gt.zarr"} ] }, classes=classes, input_arrays=input_arrays, target_arrays=target_arrays ) ``` ## Advanced Features ### Spatial Transformations Comprehensive augmentation pipeline for robust training: ```python spatial_transforms = { "mirror": { "axes": {"x": 0.5, "y": 0.5, "z": 0.1} # Probability per axis }, "rotate": { "axes": {"z": [-45, 45], "y": [-15, 15]} # Angle ranges }, "transpose": { "axes": ["x", "y"] # Axes to randomly reorder } } ``` ### Value Transformations Built-in preprocessing and augmentation transforms: ```python from cellmap_data.transforms import ( GaussianNoise, RandomContrast, RandomGamma, Binarize, NaNtoNum, GaussianBlur ) # Input preprocessing raw_transforms = T.Compose([ T.ToDtype(torch.float, scale=True), # Normalize to [0,1] GaussianNoise(std=0.1), # Add noise RandomContrast((0.8, 1.2)), # Vary contrast NaNtoNum({"nan": 0}) # Handle NaN values ]) # Target preprocessing target_transforms = T.Compose([ Binarize(threshold=0.5), # Convert to binary T.ToDtype(torch.float32) # Ensure float32 ]) ``` ### Class Relationship Handling Support for mutually exclusive classes and true negative inference: ```python # Define class relationships class_relation_dict = { "mitochondria": ["cytoplasm", "nucleus"], # Mutually exclusive "endoplasmic_reticulum": ["mitochondria"], # Cannot overlap } dataset = CellMapDataset( # ... other parameters ... classes=["mitochondria", "er", "nucleus", "cytoplasm"], class_relation_dict=class_relation_dict, # True negatives automatically inferred from relationships ) ``` ### Memory-Efficient Large Dataset Handling For datasets larger than available memory: ```python # Use subset sampling for large datasets loader = CellMapDataLoader( large_dataset, batch_size=8, iterations_per_epoch=5000, # Subsample each epoch weighted_sampler=True ) # Refresh sampler between epochs for epoch in range(num_epochs): loader.refresh() # New random subset for batch in loader: # Training code ... ``` ### Writing Predictions Generate predictions and write to disk efficiently: ```python from cellmap_data import CellMapDatasetWriter writer = CellMapDatasetWriter( raw_path="input.zarr", target_path="predictions.zarr", classes=["class1", "class2"], input_arrays=input_arrays, target_arrays=target_arrays, target_bounds={"array": {"x": [0, 1000], "y": [0, 1000], "z": [0, 100]}} ) # Write predictions tile by tile for idx in range(len(writer)): inputs = writer[idx] predictions = model(inputs) writer[idx] = {"segmentation": predictions} ``` ## Data Format Support ### Input Formats - **OME-NGFF/Zarr**: Primary format with multiscale support and full read/write capabilities - **Local/S3/GCS**: Various storage backends via TensorStore ### Multiscale Support Automatic handling of multiscale datasets: ```python # Automatically selects appropriate scale level dataset = CellMapDataset( raw_path="data.zarr", # Contains s0, s1, s2, ... scale levels target_path="labels.zarr", # ... other parameters ... ) # Multiscale input arrays can be specified input_arrays = { "raw_4nm": { "shape": (128, 128, 128), "scale": (4, 4, 4), }, "raw_8nm": { "shape": (64, 64, 64), "scale": (8, 8, 8), } } ``` ## Windows Compatibility CellMap-Data includes specific hardening for Windows to prevent native hard-crashes caused by concurrent TensorStore reads from multiple threads. ### TensorStore Read Limiter On Windows, concurrent materializations of TensorStore-backed xarray arrays (triggered by `source[center]`, `.interp`, `.__array__`, etc.) can cause the Python process to abort. A global semaphore serializes these reads automatically: ```python # The limiter activates automatically on Windows with the default TensorStore backend. # No code changes required — it is transparent to all callers. # Override the concurrency limit (default is 1 on Windows): import os os.environ["CELLMAP_MAX_CONCURRENT_READS"] = "2" # set BEFORE importing cellmap_data from cellmap_data import CellMapDataset ``` ### Environment Variables | Variable | Default | Description | |---|---|---| | `CELLMAP_DATA_BACKEND` | `"tensorstore"` | Backend for array reads (`"tensorstore"` or `"dask"`) | | `CELLMAP_MAX_WORKERS` | `8` | Max threads in the internal `ThreadPoolExecutor` | | `CELLMAP_MAX_CONCURRENT_READS` | `1` (Windows) / unlimited | Max concurrent TensorStore reads (Windows+TensorStore only) | ### Recommendations for Windows - Keep the default `num_workers=0` in `CellMapDataLoader` (safest on Windows); the internal executor still parallelizes per-array I/O within each `__getitem__` call. - If you need `num_workers > 0`, each DataLoader worker process gets its own dataset copy and its own read semaphore — this is safe. - Do **not** share a single `CellMapDataset` instance across multiple threads that each call `__getitem__` concurrently. Use separate dataset instances instead (which is exactly what DataLoader workers do). ### Explicit Shutdown `CellMapDataset` registers an `atexit` handler and exposes an explicit `close()` method for deterministic cleanup: ```python dataset = CellMapDataset(...) try: # ... training ... finally: dataset.close() # shuts down the internal ThreadPoolExecutor immediately ``` ## Performance Optimization ### Memory Management - Efficient tensor operations with minimal copying - Automatic GPU memory management - Streaming data loading for large volumes ### Parallel Processing - Multi-threaded data loading via persistent `ThreadPoolExecutor` - CUDA streams for GPU optimization - Process-safe dataset pickling ### Caching Strategy - Persistent `ThreadPoolExecutor` per process (lazy-initialized, PID-tracked) - Optimized coordinate transformations - Minimal redundant computations ## Use Cases ### 1. Cell Segmentation Training ```python # Multi-class cell segmentation classes = ["cell_boundary", "mitochondria", "nucleus", "er"] spatial_transforms = { "mirror": {"axes": {"x": 0.5, "y": 0.5}}, "rotate": {"axes": {"z": [-180, 180]}} } dataset = CellMapDataset( raw_path="em_data.zarr", target_path="segmentation_labels.zarr", classes=classes, input_arrays={"em": {"shape": (128, 128, 128), "scale": (4, 4, 4)}}, target_arrays={"labels": {"shape": (128, 128, 128), "scale": (4, 4, 4)}}, spatial_transforms=spatial_transforms, is_train=True ) ``` ### 2. Large-Scale Multi-Dataset Training ```python # Training across multiple biological samples datasplit = CellMapDataSplit( csv_path="multi_sample_split.csv", classes=organelle_classes, input_arrays=input_config, target_arrays=target_config, spatial_transforms=augmentation_config ) # Balanced sampling across datasets train_loader = CellMapDataLoader( datasplit.train_datasets_combined, batch_size=16, weighted_sampler=True, num_workers=16 ) ``` ### 3. Inference and Prediction Writing ```python # Generate predictions on new data writer = CellMapDatasetWriter( raw_path="new_sample.zarr", target_path="predictions.zarr", classes=trained_classes, input_arrays=inference_config, target_arrays=output_config, target_bounds=volume_bounds ) # Process in tiles for idx in writer.writer_indices: # Non-overlapping tiles batch = writer[idx] with torch.no_grad(): predictions = model(batch["input"]) writer[idx] = {"segmentation": predictions} ``` ## Best Practices ### Dataset Configuration - Choose patch sizes that fit comfortably in GPU memory - Enable padding for datasets smaller than patch size ### Training Optimization - Use weighted sampling for imbalanced datasets - Configure appropriate number of workers (typically 2x CPU cores) - Enable CUDA streams for multi-GPU setups ### Memory Optimization - Monitor memory usage with large datasets - Use iterations_per_epoch for very large datasets - Refresh samplers between epochs for dataset variety ### Debugging - Start with small patch sizes and single workers - Use force_has_data=True for testing with empty datasets - Check dataset.verify() before training ## API Reference For complete API documentation, visit: [https://janelia-cellmap.github.io/cellmap-data/](https://janelia-cellmap.github.io/cellmap-data/) ## Contributing We welcome contributions! Please see our [contributing guidelines](CONTRIBUTING.md) for details on: - Code style and standards - Testing requirements - Documentation expectations - Pull request process ## Citation If you use CellMap-Data in your research, please cite: ```bibtex @software{cellmap_data, title={CellMap-Data: PyTorch Data Loading for Biological Imaging}, author={Rhoades, Jeff and the CellMap Team}, url={https://github.com/janelia-cellmap/cellmap-data}, year={2024} } ``` ## License This project is licensed under the BSD 3-Clause License - see the [LICENSE](LICENSE) file for details. ## Support - 📖 [Documentation](https://janelia-cellmap.github.io/cellmap-data/) - 🐛 [Issue Tracker](https://github.com/janelia-cellmap/cellmap-data/issues) - 💬 [Discussions](https://github.com/janelia-cellmap/cellmap-data/discussions) - 📧 Contact: [rhoadesj@hhmi.org](mailto:rhoadesj@hhmi.org)
text/markdown
null
Jeff Rhoades <rhoadesj@hhmi.org>
null
null
BSD 3-Clause License
null
[ "Development Status :: 3 - Alpha", "License :: OSI Approved :: BSD License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Typing :: Typed" ]
[]
null
null
>=3.11
[]
[]
[]
[ "fsspec[http,s3]", "h5py", "ipython", "matplotlib", "neuroglancer", "numpy", "pydantic-ome-ngff", "scipy", "tensorstore", "torch", "torchvision", "tqdm", "universal-pathlib>=0.2.0", "xarray-ome-ngff", "xarray-tensorstore==0.1.5", "black; extra == \"all\"", "hatch; extra == \"all\"", "ipython; extra == \"all\"", "jupyter; extra == \"all\"", "mypy; extra == \"all\"", "pdbpp; extra == \"all\"", "pre-commit; extra == \"all\"", "pytest-cov; extra == \"all\"", "pytest-timeout; extra == \"all\"", "pytest>=6.0; extra == \"all\"", "python-semantic-release; extra == \"all\"", "rich; extra == \"all\"", "ruff; extra == \"all\"", "snakeviz; extra == \"all\"", "sphinx; extra == \"all\"", "sphinx-book-theme; extra == \"all\"", "twine; extra == \"all\"", "black; extra == \"dev\"", "hatch; extra == \"dev\"", "ipython; extra == \"dev\"", "jupyter; extra == \"dev\"", "mypy; extra == \"dev\"", "pdbpp; extra == \"dev\"", "pre-commit; extra == \"dev\"", "pytest-cov; extra == \"dev\"", "pytest-timeout; extra == \"dev\"", "pytest>=6.0; extra == \"dev\"", "python-semantic-release; extra == \"dev\"", "rich; extra == \"dev\"", "ruff; extra == \"dev\"", "snakeviz; extra == \"dev\"", "sphinx; extra == \"dev\"", "sphinx-book-theme; extra == \"dev\"", "twine; extra == \"dev\"", "black; extra == \"test\"", "mypy; extra == \"test\"", "pytest-cov; extra == \"test\"", "pytest-timeout; extra == \"test\"", "pytest>=6.0; extra == \"test\"" ]
[]
[]
[]
[ "homepage, https://github.com/janelia-cellmap/cellmap-data", "repository, https://github.com/janelia-cellmap/cellmap-data" ]
twine/6.2.0 CPython/3.11.14
2026-02-20T21:59:36.441240
cellmap_data-2026.2.20.2159.tar.gz
70,808
68/02/3bdad8f02746b27beac4437d079aae0e22cb00ea9c3b5bfc726cc3c5d99e/cellmap_data-2026.2.20.2159.tar.gz
source
sdist
null
false
67fef13634dc55eb9deddf517377f443
6416cc6da41895bb066c4364af5a4b6460e069f332d3b753a821de30d2b21a78
68023bdad8f02746b27beac4437d079aae0e22cb00ea9c3b5bfc726cc3c5d99e
null
[ "LICENSE" ]
290
2.4
mindglow
0.2.1
2D/3D Image Segmentation with PyTorch
readme = { file = "README.md", content-type = "text/markdown" }
text/markdown
null
Yasin Ansari <yasinansari7171@gmail.com>
null
null
MIT
null
[ "Development Status :: 3 - Alpha", "Intended Audience :: Science/Research", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Operating System :: OS Independent", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Scientific/Engineering :: Medical Science Apps." ]
[]
null
null
>=3.8
[]
[]
[]
[ "numpy>=1.20.0", "opencv-python>=4.5.0", "tqdm>=4.64.0" ]
[]
[]
[]
[ "Homepage, https://github.com/CallmeYasin/mindglow", "Repository, https://github.com/CallmeYasin/mindglow", "Documentation, https://mindglow.readthedocs.io", "BugTracker, https://github.com/CallmeYasin/mindglow/issues" ]
twine/6.2.0 CPython/3.12.7
2026-02-20T21:59:35.386002
mindglow-0.2.1.tar.gz
16,053
77/21/831b28998a4d60b7d5aa7e900e9fc5093b8aa77728346911bc57d3ef1c69/mindglow-0.2.1.tar.gz
source
sdist
null
false
798652a2957abe25fff56c530a51e470
6aaab3473fcc5218546cc37e73b728cc716ac014a9efd66827d40eac61017d5a
7721831b28998a4d60b7d5aa7e900e9fc5093b8aa77728346911bc57d3ef1c69
null
[]
193
2.4
orcheo
0.20.3
Add your description here
# Orcheo [![CI](https://github.com/ShaojieJiang/orcheo/actions/workflows/ci.yml/badge.svg?event=push)](https://github.com/ShaojieJiang/orcheo/actions/workflows/ci.yml?query=branch%3Amain) [![Coverage](https://coverage-badge.samuelcolvin.workers.dev/ShaojieJiang/orcheo.svg)](https://coverage-badge.samuelcolvin.workers.dev/redirect/ShaojieJiang/orcheo) [![PyPI - Core](https://img.shields.io/pypi/v/orcheo.svg?logo=python&label=core)](https://pypi.org/project/orcheo/) [![PyPI - Backend](https://img.shields.io/pypi/v/orcheo-backend.svg?logo=python&label=backend)](https://pypi.org/project/orcheo-backend/) [![PyPI - SDK](https://img.shields.io/pypi/v/orcheo-sdk.svg?logo=python&label=sdk)](https://pypi.org/project/orcheo-sdk/) [![PyPI - Agentensor](https://img.shields.io/pypi/v/agentensor.svg?logo=python&label=agentensor)](https://pypi.org/project/agentensor/) [![npm - Canvas](https://img.shields.io/npm/v/orcheo-canvas.svg?logo=npm&label=canvas)](https://www.npmjs.com/package/orcheo-canvas) [![Documentation](https://readthedocs.org/projects/orcheo/badge/?version=latest)](https://orcheo.readthedocs.io/en/latest/) Orcheo is a workflow orchestration platform designed for vibe coding — AI coding agents like Claude Code can start services, build workflows, and deploy them for you automatically. Read the [full documentation](https://orcheo.readthedocs.io/en/latest/) for guides, API reference, and examples. > **Note:** This project is currently in Beta. Expect breaking changes as we iterate rapidly towards 1.0. > **SIGIR Reviewers:** See the **[Conversational Search Examples](https://orcheo.readthedocs.io/en/latest/examples/conversational_search/)** for step-by-step demos from basic RAG to production-ready search. ## Why Orcheo? - **Vibe-coding-first**: Already using Claude Code, Codex CLI, or Cursor? You **don't** need to learn Orcheo. Install the [agent skill](https://github.com/ShaojieJiang/agent-skills) and let your AI agent handle setup, workflow creation, and deployment. - **Python-native**: Workflows are Python code powered by LangGraph — no proprietary DSL to learn. - **Backend-first**: Run headless in production; the UI is optional. ## Prerequisites - **Docker** — for running Redis and other services - **Python 3.12+** — required for the backend - **uv** — Python package manager ([installation guide](https://docs.astral.sh/uv/getting-started/installation/)) ## Quick Start The fastest way to get started with Orcheo is through the **Agent Skill** approach — let your AI coding agent handle the setup for you. > **Note:** Most AI coding agents (Claude Code, Codex CLI, Cursor) require a paid subscription. Free alternatives may exist but have not been tested with Orcheo. ### 1. Install the Orcheo Agent Skill Add the [Orcheo agent skill](https://github.com/ShaojieJiang/agent-skills) to your AI coding agent (Claude Code, Cursor, etc.) by following the installation instructions in the repo. ### 2. Let Your Agent Do the Work Once installed, simply ask your agent to: - **Set up Orcheo**: "Set up Orcheo for local development" - **Create workflows**: "Create a workflow that monitors RSS feeds and sends Slack notifications" - **Deploy workflows**: "Deploy and schedule my workflow to run every hour" Your AI agent will automatically: - Install dependencies - Start the backend server - Create and configure workflows - Handle authentication and deployment That's it! Your agent handles the complexity while you focus on describing what you want your workflows to do. ## Guides - **[Manual Setup Guide](https://orcheo.readthedocs.io/en/latest/manual_setup/)** — Installation, CLI reference, authentication, and Canvas setup - **[Conversational Search Examples](https://orcheo.readthedocs.io/en/latest/examples/conversational_search/)** — Step-by-step demos from basic RAG to production-ready search ```bash # Quick start: Run Demo 1 (no external services required) uv sync --group examples orcheo credential create openai_api_key --secret sk-your-key python examples/conversational_search/demo_2_basic_rag/demo_2.py ``` ## Reference - **[SDK Reference](https://orcheo.readthedocs.io/en/latest/sdk_reference/)** — Python SDK for programmatic workflow execution - **[Environment Variables](https://orcheo.readthedocs.io/en/latest/environment_variables/)** — Complete configuration reference ## For Developers - **[Developer Guide](https://orcheo.readthedocs.io/en/latest/manual_setup/#developer-guide)** — Repository layout, development environment, and custom nodes - **[Deployment Guide](https://orcheo.readthedocs.io/en/latest/deployment/)** — Docker Compose and PostgreSQL deployment recipes - **[Custom Nodes and Tools](https://orcheo.readthedocs.io/en/latest/custom_nodes_and_tools/)** — Extend Orcheo with your own integrations ## Contributing We welcome contributions from the community: - **Share your extensions**: Custom nodes, agent tools, and workflows that extend Orcheo's capabilities. See the [Custom Nodes and Tools guide](https://orcheo.readthedocs.io/en/latest/custom_nodes_and_tools/) for how to create and load custom extensions. - **How to contribute**: Open an [issue](https://github.com/ShaojieJiang/orcheo/issues), submit a [pull request](https://github.com/ShaojieJiang/orcheo/pulls), or start a [discussion](https://github.com/ShaojieJiang/orcheo/discussions). You can also publish and share your extensions independently for others to install. ## Citation If you use Orcheo in your research, please cite it as: ```bibtex @article{jiang2026orcheo, author = {Jiang, Shaojie and Vakulenko, Svitlana and de Rijke, Maarten}, title = {Orcheo: A Modular Full-Stack Platform for Conversational Search}, journal = {arXiv preprint arXiv:2602.14710}, year = {2026} } ```
text/markdown
null
null
null
null
null
null
[]
[]
null
null
>=3.12
[]
[]
[]
[ "agentensor>=0.0.4", "aiosqlite>=0.22.0", "croniter>=2.0.0", "cryptography>=43.0.1", "dynaconf>=3.2.4", "fastapi>=0.104.0", "fastmcp>=2.13.0.2", "feedparser>=6.0.11", "httpx>=0.28.1", "langchain-community>=0.0.10", "langchain-google-genai>=3.0.0", "langchain-mcp-adapters>=0.2.1", "langchain-openai>=0.0.5", "langchain>=1.1.3", "langgraph-checkpoint-postgres>=3.0.0", "langgraph-checkpoint-sqlite>=3.0.0", "langgraph>=1.0.5", "mcp[cli]>=1.12.0", "motor>=3.6.0", "openai-chatkit>=1.4.0", "openai>=1.0.0", "opentelemetry-api>=1.26.0", "opentelemetry-exporter-otlp>=1.26.0", "opentelemetry-sdk>=1.26.0", "pinecone-text>=0.11.0", "pinecone>=8.0.0", "psycopg[binary,pool]>=3.2.0", "py-mini-racer>=0.6.0", "pycryptodome>=3.20.0", "pydantic>=2.4.2", "pymongo>=4.13.2", "python-dotenv>=1.0.0", "python-telegram-bot>=22.0", "restrictedpython>=7.2", "rouge-score>=0.1.2", "sacrebleu>=2.3.0", "selenium>=4.32.0", "structlog>=24.4.0", "uvicorn>=0.24.0", "websockets>=12.0" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:59:11.846529
orcheo-0.20.3.tar.gz
1,821,175
48/5a/aad269a13ce18600fad22aceebed9631910e7010d67ee75f6830e5684dbe/orcheo-0.20.3.tar.gz
source
sdist
null
false
dfa8d22e00da0e1d59fe9eca1df34cc7
5ac791d300df3efac7fbad5b588311fa3a350959e6ddea657b20514fd2de3d2d
485aaad269a13ce18600fad22aceebed9631910e7010d67ee75f6830e5684dbe
null
[ "LICENSE" ]
201
2.4
truthcheck
0.4.0
Open source AI content verification
# TruthCheck 🔍 **Open source AI content verification.** Score claims 0-100 and trace their origins. [![PyPI](https://img.shields.io/pypi/v/truthcheck.svg)](https://pypi.org/project/truthcheck/) [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/licenses/MIT) ## Why? AI chatbots present web content as fact. Bad actors exploit this — a BBC journalist [showed](https://www.bbc.com/future/article/20260218-i-hacked-chatgpt-and-googles-ai-and-it-only-took-20-minutes) he could make ChatGPT call him "the best tech journalist at eating hot dogs" with one fake article. TruthCheck catches this: ```bash $ truthcheck verify "Thomas Germain is the best tech journalist at eating hot dogs" --llm gemini TruthScore: 0/100 (FALSE) ⚠️ ZERO FLAG: Content identified as satire ``` ## Install ```bash pip install truthcheck ``` ## Setup **Search:** Works out of the box with DuckDuckGo (free, no key needed). **LLM (required for `--llm` flag):** ```bash export GOOGLE_API_KEY=AIza... # Gemini (recommended) # or OPENAI_API_KEY, ANTHROPIC_API_KEY ``` ## Usage ```bash # Verify a claim (with LLM for deep analysis) truthcheck verify "Some claim" --llm gemini # Trace claim origin truthcheck trace "Some claim" # Check URL(s) — single or multiple truthcheck check https://reuters.com truthcheck check "Sources: https://reuters.com and https://bbc.com" # Check publisher reputation truthcheck lookup breitbart.com ``` ### Check URLs Verify URLs and detect hallucinations (URLs that don't exist): ```bash $ truthcheck check "Sources: https://reuters.com/fake-article and https://bbc.com" URLs Found: 2 Summary: 🚨 1 broken/hallucinated URL(s) | ✅ 1 sources verified 🚨 Broken/Hallucinated URLs: ✗ https://reuters.com/fake-article (404 Not Found) Verified URLs: 81% https://bbc.com ``` Also works with files: ```bash truthcheck check -f response.txt ``` ### Trace Example ```bash $ truthcheck trace "Thomas Germain is the best tech journalist at eating hot dogs" 🎯 ORIGIN Domain: tomgermain.com Date: 2026-02-05 📅 TIMELINE 🥇 [2026-02-05] tomgermain.com 🥈 [2026-02-18] bbc.com 🥉 [2026-02-18] gizmodo.com 📊 STATS Sources: 14 | Date range: 2026-02-05 → 2026-02-19 ``` > 📋 **More examples:** See [TEST_CASES.md](docs/TEST_CASES.md) for detailed verification results on common misinformation claims. ### Python ```python from truthcheck import verify_claim, trace_claim from truthcheck.search import DuckDuckGoProvider result = verify_claim("Earth is flat", search_provider=DuckDuckGoProvider()) print(f"TruthScore: {result.truthscore}/100") origin = trace_claim("Some viral claim") print(f"Origin: {origin['origin']['domain']}") ``` ## How It Works TruthScore weighs four factors equally: | Factor | Weight | |--------|--------| | Publisher credibility (origin) | 25% | | Content analysis | 25% | | Corroboration | 25% | | Fact-checker verdicts | 25% | Fact-checkers are weighted by their own MBFC credibility rating (reduces bias from any single source). **Zero flags** (satire, self-published, fake experiments) force score to 0. ## MCP Server Works with Claude Desktop and Cursor: ```json { "mcpServers": { "truthcheck": { "command": "truthcheck-mcp" } } } ``` ## License MIT --- [Issues](https://github.com/baiyishr/truthcheck/issues) · [Docs](https://github.com/baiyishr/truthcheck/tree/main/docs) · [Test Cases](docs/TEST_CASES.md)
text/markdown
TruthCheck Contributors
null
null
null
null
ai, fact-check, mcp, misinformation, verification
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Topic :: Scientific/Engineering :: Artificial Intelligence" ]
[]
null
null
>=3.10
[]
[]
[]
[ "click>=8.0", "datasketch>=1.6", "ddgs>=7.0", "mcp>=1.0", "pyyaml>=6.0", "requests>=2.28", "anthropic>=0.18; extra == \"anthropic\"", "anthropic>=0.18; extra == \"dev\"", "black>=23.0; extra == \"dev\"", "google-genai>=1.0; extra == \"dev\"", "mypy>=1.0; extra == \"dev\"", "openai>=1.0; extra == \"dev\"", "pytest-cov>=4.0; extra == \"dev\"", "pytest-mock>=3.10; extra == \"dev\"", "pytest>=7.0; extra == \"dev\"", "python-dotenv>=1.0; extra == \"dev\"", "ruff>=0.1; extra == \"dev\"", "python-dotenv>=1.0; extra == \"dotenv\"", "google-genai>=1.0; extra == \"gemini\"", "anthropic>=0.18; extra == \"llm\"", "google-genai>=1.0; extra == \"llm\"", "openai>=1.0; extra == \"llm\"", "openai>=1.0; extra == \"openai\"" ]
[]
[]
[]
[ "Homepage, https://github.com/baiyishr/truthcheck", "Documentation, https://github.com/baiyishr/truthcheck#readme", "Repository, https://github.com/baiyishr/truthcheck", "Issues, https://github.com/baiyishr/truthcheck/issues" ]
twine/6.2.0 CPython/3.14.2
2026-02-20T21:59:10.691638
truthcheck-0.4.0.tar.gz
290,116
0e/26/a02e08488ef9d049a108ba6f6e0d485cfcc1e55a4910d1cfe9ae7f619538/truthcheck-0.4.0.tar.gz
source
sdist
null
false
e905fb57426e573c21def3abc17e4163
9d67bb5813f1f435fcaa1f2cc27420894000d7aa39052e7661037d120e596d73
0e26a02e08488ef9d049a108ba6f6e0d485cfcc1e55a4910d1cfe9ae7f619538
MIT
[]
185
2.4
NonlinearTMM
1.4.1
Nonlinear transfer matrix method
[![PyPI version](https://badge.fury.io/py/NonlinearTMM.svg)](https://badge.fury.io/py/NonlinearTMM) [![Python](https://img.shields.io/pypi/pyversions/NonlinearTMM)](https://pypi.org/project/NonlinearTMM/) [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/licenses/MIT) [![Pytest](https://github.com/ardiloot/NonlinearTMM/actions/workflows/pytest.yml/badge.svg)](https://github.com/ardiloot/NonlinearTMM/actions/workflows/pytest.yml) [![Pre-commit](https://github.com/ardiloot/NonlinearTMM/actions/workflows/pre-commit.yml/badge.svg)](https://github.com/ardiloot/NonlinearTMM/actions/workflows/pre-commit.yml) [![Build and upload to PyPI](https://github.com/ardiloot/NonlinearTMM/actions/workflows/publish.yml/badge.svg)](https://github.com/ardiloot/NonlinearTMM/actions/workflows/publish.yml) # NonlinearTMM: Nonlinear Transfer-Matrix Method A Python library for optical simulations of **multilayer structures** using the transfer-matrix method, extended to support **nonlinear processes** (SHG, SFG, DFG) and **Gaussian beam propagation**. <p align="center"> <img src="docs/images/TMMForWaves-example.png" alt="Gaussian beam exciting surface plasmon polaritons" width="700"> </p> > **See also:** [GeneralTmm](https://github.com/ardiloot/GeneralTmm) — a 4×4 TMM for **anisotropic** (birefringent) multilayer structures. ## Table of Contents - [Features](#features) - [Installation](#installation) - [API Overview](#api-overview) - [Examples](#examples) - [Surface Plasmon Polaritons](#surface-plasmon-polaritons--exampletmmpy) - [Gaussian Beam Excitation](#gaussian-beam-excitation--exampletmmforwavespy) - [Second-Harmonic Generation](#second-harmonic-generation--examplesecondordernonlineartmmpy) - [References](#references) - [Documentation](#documentation) - [Development](#development) - [Setup](#setup) - [Running tests](#running-tests) - [Code formatting and linting](#code-formatting-and-linting) - [CI overview](#ci-overview) - [Releasing](#releasing) - [License](#license) ## Features - **Standard TMM** — reflection, transmission, absorption for p- and s-polarized plane waves at arbitrary angles - **Parameter sweeps** — over wavelength, angle of incidence, layer thickness, or any other parameter - **1D and 2D electromagnetic field profiles** — E and H field distributions through the structure - **Field enhancement** — calculation of field enhancement factors (e.g. for SPP excitation) - **Gaussian beam propagation** — any beam profile through layered structures, not just plane waves - **Second-order nonlinear processes** — SHG, SFG, DFG in multilayer structures - **Wavelength-dependent materials** — interpolated from measured optical data (YAML format) - **High performance** — C++ core (Eigen) with Cython bindings, OpenMP parallelization - **Cross-platform wheels** — Linux (x86_64), Windows (x64, ARM64), macOS (ARM64); Python 3.10–3.14 ## Installation ```bash pip install NonlinearTMM ``` Pre-built wheels are available for most platforms. A C++ compiler is only needed when installing from source. ## API Overview The library exposes three main classes: `Material`, `TMM`, and `SecondOrderNLTMM`. | Class / method | Purpose | |---|---| | `Material(wls, ns)` | Wavelength-dependent material from arrays of λ and complex n | | `Material.Static(n)` | Constant refractive index (shortcut) | | `TMM(wl=…, pol=…, I0=…)` | Create a solver; `wl` = wavelength (m), `pol` = `"p"` or `"s"` | | `tmm.AddLayer(d, mat)` | Append layer (`d` in m, `inf` for semi-infinite) | | `tmm.Sweep(param, values)` | Solve for an array of values of any parameter | | `tmm.GetFields(zs)` | E, H field profiles along the layer normal | | `tmm.GetFields2D(zs, xs)` | E, H on a 2-D grid | | `tmm.GetEnhancement(layerNr)` | Field enhancement in a given layer | | `tmm.wave` | Access `_Wave` parameters for Gaussian beam calculations | | `tmm.WaveSweep(param, values)` | Parameter sweep for beam calculations | | `tmm.WaveGetFields2D(zs, xs)` | 2-D field map for beam excitation | | `SecondOrderNLTMM(…)` | Second-order nonlinear TMM (SHG, SFG, DFG) | For the full API, see the [reference documentation](https://ardiloot.github.io/NonlinearTMM/Reference.html). ## Examples ### Surface Plasmon Polaritons — [ExampleTMM.py](Examples/ExampleTMM.py) Kretschmann configuration (prism | 50 nm Ag | air) at 532 nm. Demonstrates reflection sweeps, field enhancement, and 1D/2D field visualization of surface plasmon polaritons. ```python import math import numpy as np from NonlinearTMM import TMM, Material # Materials prism = Material.Static(1.5) ag = Material.Static(0.054007 + 3.4290j) # Silver @ 532nm air = Material.Static(1.0) # Set up TMM (Kretschmann configuration) tmm = TMM(wl=532e-9, pol="p", I0=1.0) tmm.AddLayer(math.inf, prism) tmm.AddLayer(50e-9, ag) tmm.AddLayer(math.inf, air) # Sweep angle of incidence betas = np.sin(np.radians(np.linspace(0, 80, 500))) * 1.5 result = tmm.Sweep("beta", betas, outEnh=True, layerNr=2) ``` <p align="center"> <img src="docs/images/TMM-example.png" alt="SPP reflection, enhancement, and field profiles" width="700"> </p> ### Gaussian Beam Excitation — [ExampleTMMForWaves.py](Examples/ExampleTMMForWaves.py) The same Kretschmann structure excited by a 10 mW Gaussian beam (waist 10 μm). Shows how finite beam width affects resonance depth and field enhancement. <p align="center"> <img src="docs/images/TMMForWaves-example.png" alt="Gaussian beam SPP excitation" width="700"> </p> ### Second-Harmonic Generation — [ExampleSecondOrderNonlinearTmm.py](Examples/ExampleSecondOrderNonlinearTmm.py) Second-harmonic generation (SHG) in a 1 mm nonlinear crystal with χ⁽²⁾ nonlinearity. Two s-polarized pump beams at 1000 nm generate a second-harmonic signal at 500 nm. The `SecondOrderNLTMM` class also supports sum-frequency generation (SFG) and difference-frequency generation (DFG). <p align="center"> <img src="docs/images/SecondOrderNLTMM-example.png" alt="SHG reflected and transmitted intensity vs beta" width="700"> </p> ## References > Loot, A., & Hizhnyakov, V. (2017). Extension of standard transfer-matrix method for three-wave mixing for plasmonic structures. *Applied Physics A*, 123(3), 152. [doi:10.1007/s00339-016-0733-0](https://link.springer.com/article/10.1007%2Fs00339-016-0733-0) > > Loot, A., & Hizhnyakov, V. (2018). Modeling of enhanced spontaneous parametric down-conversion in plasmonic and dielectric structures with realistic waves. *Journal of Optics*, 20, 055502. [doi:10.1088/2040-8986/aab6c0](https://doi.org/10.1088/2040-8986/aab6c0) ## Documentation Full documentation is available at https://ardiloot.github.io/NonlinearTMM/. - [Getting started](https://ardiloot.github.io/NonlinearTMM/GettingStarted.html) — installation, package structure, examples - [API reference](https://ardiloot.github.io/NonlinearTMM/Reference.html) — complete class and method reference ## Development ### Setup ```bash git clone --recurse-submodules https://github.com/ardiloot/NonlinearTMM.git cd NonlinearTMM # Install uv if not already installed: # https://docs.astral.sh/uv/getting-started/installation/ # Create venv, build the C++ extension, and install all dependencies uv sync ``` ### Running tests ```bash uv run pytest -v ``` ### Code formatting and linting [Pre-commit](https://pre-commit.com/) hooks are configured to enforce formatting (ruff, clang-format) and catch common issues. To install the git hook locally: ```bash uv run pre-commit install ``` To run all checks manually: ```bash uv run pre-commit run --all-files ``` ### CI overview | Workflow | Trigger | What it does | |----------|---------|--------------| | [Pytest](.github/workflows/pytest.yml) | Push to `master` / PRs | Tests on {ubuntu, windows, macos} × Python {3.10–3.14} | | [Pre-commit](.github/workflows/pre-commit.yml) | Push to `master` / PRs | Runs ruff, clang-format, ty, and other checks | | [Publish to PyPI](.github/workflows/publish.yml) | Release published | Builds wheels + sdist via cibuildwheel, uploads to PyPI | | [Publish docs](.github/workflows/publish_docs.yml) | Release published | Builds Sphinx docs and deploys to GitHub Pages | ## Releasing Versioning is handled automatically by [setuptools-scm](https://github.com/pypa/setuptools-scm) from git tags. 1. **Ensure CI is green** on the `master` branch. 2. **Create a new release** on GitHub: - Go to [Releases](https://github.com/ardiloot/NonlinearTMM/releases) → **Draft a new release** - Create a new tag following [PEP 440](https://peps.python.org/pep-0440/) (e.g. `v1.2.0`) - Target the `master` branch (or a specific commit on master) - Click **Generate release notes** for an auto-generated changelog - For pre-releases (e.g. `v1.2.0rc1`), check **Set as a pre-release** — these upload to TestPyPI instead of PyPI 3. **Publish the release** — the workflow builds wheels for Linux (x86_64), Windows (x64, ARM64), and macOS (ARM64), and uploads to [PyPI](https://pypi.org/project/NonlinearTMM/). ## License [MIT](LICENSE)
text/markdown
null
Ardi Loot <ardi.loot@outlook.com>
null
null
null
null
[ "Development Status :: 5 - Production/Stable", "Intended Audience :: Science/Research", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Topic :: Scientific/Engineering :: Physics" ]
[]
null
null
>=3.10
[]
[]
[]
[ "numpy>=2", "scipy>=1.14", "eigency>=3.4.0.0" ]
[]
[]
[]
[ "Homepage, https://github.com/ardiloot/NonlinearTMM", "Repository, https://github.com/ardiloot/NonlinearTMM", "Documentation, https://ardiloot.github.io/NonlinearTMM/" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:58:48.838675
nonlineartmm-1.4.1.tar.gz
2,283,183
93/8c/3052c61e9a8f12d7b85cb4d052813ce2404e0614da25eafdbab57847c39d/nonlineartmm-1.4.1.tar.gz
source
sdist
null
false
f3bc3033f9c5f39e8c8eb801eb46bda3
70f2e5a58a1f4fa25bdcc6e68fd6204fae926d3e521250ce05b14bf2decede1a
938c3052c61e9a8f12d7b85cb4d052813ce2404e0614da25eafdbab57847c39d
MIT
[ "LICENSE" ]
0
2.4
auxjad
1.0.6
Auxiliary classes and functions for Abjad.
|Auxjad image| |PyPI| |Build| |Python versions| |License| |Bug report| |Documentation| Auxjad is a library of auxiliary classes and functions for `Abjad 3.4`_ aimed at composers of algorithmic music. All classes and functions have a ``__doc__`` attribute with usage instructions. Documentation is available at the `Auxjad Docs`_ webpage. Bugs can be reported through the project's `Issue Tracker`_. This library is published under the `MIT License`_. Installation ============ The recommended way to install Auxjad is via `pip`_:: ~$ pip install --user auxjad If you are using virtual environments, simply use:: ~$ pip install auxjad Auxjad requires `Python 3.10`_ and `LilyPond 2.24`_ or later, as well as `Abjad 3.4`_. Please note that Auxjad is **not compatible** with newever versions of Abjad. .. |Auxjad image| image:: https://raw.githubusercontent.com/gilbertohasnofb/auxjad/master/assets/auxjad-banner.png :target: https://github.com/gilbertohasnofb/auxjad .. |PyPI| image:: https://img.shields.io/pypi/v/auxjad.svg?style=for-the-badge :target: https://pypi.python.org/pypi/auxjad .. |Build| image:: https://img.shields.io/github/actions/workflow/status/gilbertohasnofb/auxjad/github-actions.yml?style=for-the-badge :target: https://github.com/gilbertohasnofb/auxjad/actions/workflows/github-actions.yml .. |Python versions| image:: https://img.shields.io/pypi/pyversions/auxjad.svg?style=for-the-badge :target: https://www.python.org/downloads/release/python-3100/ .. |License| image:: https://img.shields.io/badge/license-MIT-blue?style=for-the-badge :target: https://github.com/gilbertohasnofb/auxjad/blob/master/LICENSE .. |Bug report| image:: https://img.shields.io/badge/bug-report-red.svg?style=for-the-badge :target: https://github.com/gilbertohasnofb/auxjad/issues .. |Documentation| image:: https://img.shields.io/badge/docs-auxjad.docs-yellow?style=for-the-badge :target: https://gilbertohasnofb.github.io/auxjad-docs/ .. _`Auxjad Docs`: https://gilbertohasnofb.github.io/auxjad-docs/ .. _`Issue Tracker`: https://github.com/gilbertohasnofb/auxjad/issues .. _`MIT License`: https://github.com/gilbertohasnofb/auxjad/blob/master/LICENSE .. _pip: https://pip.pypa.io/en/stable/ .. _`Python 3.10`: https://www.python.org/ .. _`Abjad 3.4`: https://abjad.github.io/ .. _`LilyPond 2.24`: http://lilypond.org/
text/x-rst
null
Gilberto Agostinho <gilbertohasnofb@gmail.com>
null
null
MIT
auxjad, algorithmic composition, generative music, computer music, music composition, music notation, lilypond, abjad
[ "Development Status :: 5 - Production/Stable", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Artistic Software", "Topic :: Utilities" ]
[]
null
null
>=3.10
[]
[]
[]
[ "abjad==3.4", "black>=25.1.0", "flake8>=7.2.0", "isort>=6.0.1", "pydocstyle>=6.3.0", "pygments>=2.19.2", "pytest>=8.4.0", "setuptools>=80.9.0", "sphinx>=7.4.0" ]
[]
[]
[]
[ "homepage, https://gilbertohasnofb.github.io/auxjad-docs/", "source, https://github.com/gilbertohasnofb/auxjad" ]
twine/6.2.0 CPython/3.12.3
2026-02-20T21:58:44.751704
auxjad-1.0.6.tar.gz
448,518
31/50/5f179886df70b9faa02733c40e1a8a1ea70055a42a605e0c2cb79c45d42d/auxjad-1.0.6.tar.gz
source
sdist
null
false
af5918380d78b3f0506790642a6cd735
63dcc3370cdcb7d6d417d1c35c88716e4ca1f27ed31d6ffc1422fe691f0378d7
31505f179886df70b9faa02733c40e1a8a1ea70055a42a605e0c2cb79c45d42d
null
[ "LICENSE" ]
151
2.4
getv
0.2.10
Universal .env variable manager — read, write, encrypt, delegate across services and devices
# getv — Universal .env Variable Manager [![PyPI version](https://badge.fury.io/py/getv.svg)](https://badge.fury.io/py/getv) [![Python versions](https://img.shields.io/pypi/pyversions/getv)](https://pypi.org/project/getv/) [![License: Apache-2.0](https://img.shields.io/badge/License-Apache%202.0-blue.svg)](https://opensource.org/licenses/Apache-2.0) [![Downloads](https://img.shields.io/pypi/dm/getv)](https://pypi.org/project/getv/) [![Tests](https://github.com/wronai/getv/workflows/Tests/badge.svg)](https://github.com/wronai/getv/actions) We increasingly work with APIs, especially in the context of LLMs, constantly needing to provide new keys and then copy them to other tools, `.env` files. What if when copying to the clipboard, they were automatically saved to `.env` from which various applications like ollama or libraries like liteLLM could retrieve the specified `API TOKEN`? - copy an `API KEY` in your browser - run: `getv grab` - use multiple keys directly: `getv exec llm groq -- python app.py` Read, write, encrypt, and delegate environment variables across services and devices. ![img.png](img.png) Copy to the clipboard and run `getv grab` to detect and save the API key ```bash $ getv grab Detected: groq (GROQ_API_KEY) Key: gsk_Y1xV...TNpA Source: Prefix match Domain: console.groq.com Category: llm Profile: ~/.getv/llm/groq.env Saved to /home/tom/.getv/llm/groq.env Usage: getv get llm groq GROQ_API_KEY getv exec llm groq -- python app.py ``` without any plugins, managers, or integrations... Clipboard → .env Auto-Detection No other CLI tool has this feature. Copy any API key and run getv grab — it auto-detects the provider, saves to the right profile, and shows usage commands. Perfect for rapid RPi + LLM development workflow. ```bash $ getv grab # copy API key → auto-detect provider → save $ getv exec llm groq -- python app.py # run with profile injected $ getv ssh rpi3 "uname -a" # SSH using stored profile ``` ![getv.gif](getv.gif) ## Why getv? - **Clipboard → .env** — `getv grab` auto-detects 19 API key prefixes from clipboard - **Built-in integrations** — SSH, curl, Docker, LiteLLM, Ollama, Pydantic — no plugins - **Smart profiles** — organize by category (`llm/`, `devices/`, `tokens/`) with per-app defaults - **One-liner power** — process substitution, pipes, shell eval ## Install ```bash pip install getv # core pip install "getv[crypto]" # + encryption (Fernet) pip install "getv[all]" # everything ``` ## Quick Start ```bash # Save a profile getv set llm groq LLM_MODEL=groq/llama-3.3-70b-versatile GROQ_API_KEY=gsk_xxx # Read a variable getv get llm groq LLM_MODEL # List profiles (secrets masked) getv list llm # Run with env injected getv exec llm groq -- python my_script.py # Auto-detect API key from clipboard getv grab # Export formats getv export llm groq --format json getv export llm groq --format shell getv export llm groq --format docker ``` ### Python API ```python from getv import EnvStore, ProfileManager # Single .env file store = EnvStore("~/.myapp/.env") store.set("DB_HOST", "localhost").set("DB_PORT", "5432").save() # Named profiles pm = ProfileManager("~/.getv") pm.add_category("llm", required_keys=["LLM_MODEL"]) pm.set("llm", "groq", {"LLM_MODEL": "groq/llama-3.3-70b-versatile", "GROQ_API_KEY": "gsk_xxx"}) # Merge profiles cfg = pm.merge_profiles({"APP_NAME": "myapp"}, llm="groq", devices="rpi3") ``` ## Profile Directory ```text ~/.getv/ ├── .fernet.key ← encryption key (chmod 600) ├── defaults/ ← per-app profile selections │ ├── fixpi.conf → llm=groq, devices=rpi3 │ └── prellm.conf → llm=openrouter ├── devices/ │ ├── rpi3.env │ └── rpi4-prod.env ├── llm/ │ ├── groq.env │ └── openrouter.env └── tokens/ └── github.env ``` ## CLI Reference | Command | Description | |---------|-------------| | `getv set CATEGORY PROFILE KEY=VAL...` | Create/update profile | | `getv get CATEGORY PROFILE KEY` | Get single value | | `getv list [CATEGORY [PROFILE]]` | List categories/profiles/vars | | `getv delete CATEGORY PROFILE` | Delete profile | | `getv export CATEGORY PROFILE --format FMT` | Export (json/shell/docker/env/pydantic) | | `getv encrypt CATEGORY PROFILE` | Encrypt sensitive values | | `getv decrypt CATEGORY PROFILE` | Decrypt values | | `getv exec CATEGORY PROFILE -- CMD...` | Run with profile env | | `getv use APP CATEGORY PROFILE` | Set app default | | `getv defaults [APP]` | Show app defaults | | `getv ssh PROFILE [CMD]` | SSH to device | | `getv curl PROFILE URL` | Authenticated API call | | `getv grab [--dry-run]` | Auto-detect API key from clipboard | | `getv diff CATEGORY A B` | Compare two profiles | | `getv copy CAT/SRC CAT/DST` | Clone profile | | `getv import FILE [CAT PROFILE]` | Import from .env / docker-compose | | `getv init` | Interactive setup wizard | ## Documentation | Document | Description | |----------|-------------| | [docs/INTEGRATIONS.md](docs/INTEGRATIONS.md) | SSH, LiteLLM, Ollama, Docker, curl, Pydantic, nfo, file watcher | | [docs/GRAB.md](docs/GRAB.md) | Clipboard detection — 19 prefixes, browser history, Python API | | [docs/SECURITY.md](docs/SECURITY.md) | Masking, Fernet encryption, key rotation, validation, format export | | [docs/EXAMPLES.md](docs/EXAMPLES.md) | 20+ one-liner examples — pipes, process substitution, cron, Docker | | [docs/COMPARISON.md](docs/COMPARISON.md) | getv vs direnv vs dotenvx vs envie | ## Environment Variables | Variable | Default | Description | |----------|---------|-------------| | `GETV_HOME` | `~/.getv` | Base directory for profiles | ## Adopted by - **[fixpi](https://github.com/zlecenia/c2004/tree/main/fixPI)** — SSH + LLM diagnostic agent - **[prellm](https://github.com/wronai/prellm)** — LLM preprocessing proxy - **[code2logic](https://github.com/wronai/code2logic)** — Code analysis engine - **[amen](https://github.com/wronai/amen)** — Intent-iterative AI gateway - **[marksync](https://github.com/wronai/marksync)** — Markdown sync server - **[curllm](https://github.com/wronai/curllm)** — LLM-powered web automation ## Development ```bash git clone https://github.com/wronai/getv.git cd getv python -m venv .venv && source .venv/bin/activate pip install -e ".[dev]" pytest # 190 tests ``` ## License Apache License 2.0 - see [LICENSE](LICENSE) for details. ## Author Created by **Tom Sapletta** - [tom@sapletta.com](mailto:tom@sapletta.com)
text/markdown
null
Tom Sapletta <tom@sapletta.com>
null
null
null
env, dotenv, config, secrets, encryption, profiles, cli
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Systems Administration" ]
[]
null
null
>=3.9
[]
[]
[]
[ "python-dotenv>=1.0.0", "click>=8.0", "cryptography>=41.0; extra == \"crypto\"", "pydantic>=2.0; extra == \"pydantic\"", "pydantic-settings>=2.0; extra == \"pydantic\"", "cryptography>=41.0; extra == \"all\"", "pydantic>=2.0; extra == \"all\"", "pydantic-settings>=2.0; extra == \"all\"", "pytest>=7.0; extra == \"dev\"", "pytest-cov>=4.0; extra == \"dev\"" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.13.7
2026-02-20T21:58:42.331219
getv-0.2.10.tar.gz
49,912
c9/cd/f78124ed2f992f46f73ec8c82c2389d8a9b14859e027edb43aa853c37411/getv-0.2.10.tar.gz
source
sdist
null
false
d40b0dd33c3cbf64c1bf2d5b55fc719f
1279a3909eda281e14ab7e064e64b3cc327e147902d75b30f13adb8285b57b46
c9cdf78124ed2f992f46f73ec8c82c2389d8a9b14859e027edb43aa853c37411
Apache-2.0
[ "LICENSE" ]
199
2.4
langchain-callback-parquet-logger
3.2.1
A Parquet-based callback handler for logging LangChain LLM interactions
# LangChain Parquet Logger High-performance logging for LangChain - save all your LLM interactions to Parquet files for analysis. ## Quick Start (2 minutes) ### Install ```bash pip install langchain-callback-parquet-logger # With S3 support pip install "langchain-callback-parquet-logger[s3]" ``` ### Basic Usage ```python from langchain_callback_parquet_logger import ParquetLogger from langchain_openai import ChatOpenAI # Add logger to any LangChain LLM logger = ParquetLogger("./logs") llm = ChatOpenAI(callbacks=[logger]) response = llm.invoke("What is 2+2?") # Your logs are automatically saved to ./logs/ ``` ### Batch Processing ```python import pandas as pd from langchain_callback_parquet_logger import batch_process # Your data df = pd.DataFrame({ 'prompt': ['What is AI?', 'Explain quantum computing'] }) # Process it (logs automatically saved) results = await batch_process(df) ``` That's it! Your logs are in Parquet format, ready for analysis. ## Core Features ### 1. Custom Tracking IDs Track specific requests with custom IDs and descriptions: ```python from langchain_callback_parquet_logger import ParquetLogger, with_tags logger = ParquetLogger("./logs") llm = ChatOpenAI(callbacks=[logger]) # Add custom ID with description to track this specific request response = llm.invoke( "What is quantum computing?", config=with_tags( custom_id="user-123-session-456", custom_id_description="User session from mobile app" ) ) ``` ### 2. Batch Processing (Simple) ```python import pandas as pd from langchain_openai import ChatOpenAI from langchain_callback_parquet_logger import batch_process, with_tags, LLMConfig # Prepare your data df = pd.DataFrame({ 'prompt': ['What is AI?', 'Explain DNA'], 'config': [ with_tags(custom_id='q1', custom_id_description='Science FAQ'), with_tags(custom_id='q2', custom_id_description='Science FAQ') ] }) # Process with automatic logging results = await batch_process( df, llm_config=LLMConfig( llm_class=ChatOpenAI, llm_kwargs={'model': 'gpt-4', 'temperature': 0.7} ) ) ``` ### 3. Batch Processing (Full Configuration) ```python import pandas as pd from langchain_openai import ChatOpenAI from langchain_callback_parquet_logger import ( batch_process, with_tags, LLMConfig, JobConfig, StorageConfig, ProcessingConfig, ColumnConfig, S3Config ) # Prepare your data with custom column names df = pd.DataFrame({ 'question': ['What is AI?', 'Explain DNA', 'What is quantum computing?'], 'user_id': ['user1', 'user2', 'user3'], 'tool_list': [[tool1, tool2], None, [tool3]] # Optional tools }) # Add config for each row (required) df['run_config'] = df['user_id'].apply(lambda x: with_tags( custom_id=x, tags=['production', 'v2'] )) # Process with ALL configuration options results = await batch_process( df, # LLM configuration llm_config=LLMConfig( llm_class=ChatOpenAI, llm_kwargs={'model': 'gpt-4', 'temperature': 0.7}, model_kwargs={'top_p': 0.9}, # Additional model parameters structured_output=None # or Pydantic model for structured responses ), # Job metadata configuration (all fields except category are optional) job_config=JobConfig( category="research", subcategory="science", # Optional, defaults to None description="Analyzing scientific questions", # Optional version="2.0.0", # Optional environment="production", # Optional metadata={"team": "data-science", "priority": "high"} # Optional ), # Storage configuration storage_config=StorageConfig( output_dir="./batch_logs", path_template="{job_category}/{date}/{job_subcategory}/v{job_version_safe}", # Custom path structure with version s3_config=S3Config( bucket="my-llm-logs", prefix="langchain-logs/", on_failure="continue", # or "error" to fail on S3 errors retry_attempts=3 ) ), # Processing configuration processing_config=ProcessingConfig( max_concurrency=100, # Parallel requests buffer_size=1000, # Logger buffer size show_progress=True, # Progress bar with real-time updates return_exceptions=True, # Don't fail on single errors return_results=True, # Set False for huge datasets to save memory event_types=['llm_start', 'llm_end', 'llm_error'], # Events to log partition_on="date" # Partition strategy ), # Column name configuration (if not using defaults) column_config=ColumnConfig( prompt="question", # Your prompt column name config="run_config", # Your config column name tools="tool_list" # Your tools column name (optional) ) ) # Results are returned AND saved to Parquet files df['answer'] = results ``` ### 4. S3 Upload For production and cloud environments: ```python from langchain_callback_parquet_logger import ParquetLogger, S3Config logger = ParquetLogger( log_dir="./logs", s3_config=S3Config( bucket="my-llm-logs", prefix="production/", on_failure="error" # Fail fast in production ) ) ``` ### 5. Event Type Selection Choose what events to log: ```python # Default: Only LLM events logger = ParquetLogger("./logs") # Log everything logger = ParquetLogger( "./logs", event_types=['llm_start', 'llm_end', 'llm_error', 'chain_start', 'chain_end', 'chain_error', 'tool_start', 'tool_end', 'tool_error'] ) ``` ## Reading Your Logs ```python import pandas as pd import json # Read all logs df = pd.read_parquet("./logs") # Parse the payload df['data'] = df['payload'].apply(json.loads) # Analyze token usage df['tokens'] = df['data'].apply(lambda x: x.get('data', {}).get('outputs', {}).get('usage', {}).get('total_tokens')) ``` ## v2.0 Breaking Changes If upgrading from v1.x: ### Old (v1.x) ```python logger = ParquetLogger( log_dir="./logs", s3_bucket="my-bucket", s3_prefix="logs/", s3_on_failure="error" ) ``` ### New (v2.0) ```python from langchain_callback_parquet_logger import ParquetLogger, S3Config logger = ParquetLogger( log_dir="./logs", s3_config=S3Config( bucket="my-bucket", prefix="logs/", on_failure="error" ) ) ``` ### batch_process changes: - Now uses LLMConfig dataclass for LLM configuration - Dataclass configs replace multiple parameters - Column renamed from `logger_custom_id` to `custom_id` - See batch processing examples above #### Old batch_process (v1.x) ```python await batch_process( df, llm=llm_instance, # or llm_class with llm_kwargs structured_output=MyModel ) ``` #### New batch_process (v2.0) ```python await batch_process( df, llm_config=LLMConfig( llm_class=ChatOpenAI, llm_kwargs={'model': 'gpt-4'}, model_kwargs={'top_p': 0.9}, # Additional API params structured_output=MyModel ) ) ## Configuration Classes ### ParquetLogger - `log_dir`: Where to save logs (default: "./llm_logs") - `buffer_size`: Entries before auto-flush (default: 100) - `s3_config`: Optional S3Config for uploads ### LLMConfig - `llm_class`: The LangChain LLM class to instantiate (e.g., ChatOpenAI) - `llm_kwargs`: Arguments for the LLM constructor (model, temperature, etc.) - `model_kwargs`: Additional API parameters (top_p, frequency_penalty, etc.) - `structured_output`: Optional Pydantic model for structured responses ### JobConfig - `category`: Job category (required, default: "batch_processing") - `subcategory`: Job subcategory (optional, default: None) - `version`: Version string (optional, default: None) - `environment`: Environment name (optional, default: None) - `description`: Job description (optional, default: None) - `metadata`: Additional metadata dict (optional, default: None) ### StorageConfig - `output_dir`: Local directory (default: "./batch_logs") - `path_template`: Path template for organizing files (default: "{job_category}/{job_subcategory}/v{job_version_safe}") - Available variables: `job_category`, `job_subcategory`, `job_version` (original), `job_version_safe` (dots replaced with underscores), `environment`, `date` - Example paths: `ml_training/image_classification/v2_1_0/` or `research/nlp/vunversioned/` (when no version specified) - `s3_config`: Optional S3Config for uploads ### S3Config - `bucket`: S3 bucket name - `prefix`: S3 prefix/folder (default: "langchain-logs/") - `on_failure`: "error" or "continue" (default: "error") ## Advanced Usage ### Low-Level Batch Processing If you need direct control over logging: ```python from langchain_callback_parquet_logger import batch_run, ParquetLogger # Setup your own logging with ParquetLogger('./logs') as logger: llm = ChatOpenAI(callbacks=[logger]) # Use low-level batch_run results = await batch_run(df, llm, max_concurrency=100) ``` ### Context Manager (Notebooks) For Jupyter notebooks, use context manager for immediate writes: ```python with ParquetLogger('./logs', buffer_size=1) as logger: llm = ChatOpenAI(callbacks=[logger]) response = llm.invoke("Hello!") # Logs are guaranteed to be written ``` ## Log Schema | Column | Type | Description | |--------|------|-------------| | `timestamp` | timestamp | Event time (UTC) | | `run_id` | string | Unique run ID | | `parent_run_id` | string | Parent run ID for nested calls | | `custom_id` | string | Your custom tracking ID | | `event_type` | string | Event type (llm_start, llm_end, etc.) | | `logger_metadata` | string | JSON metadata | | `payload` | string | Full event data as JSON | ## Payload Structure All events use a consistent JSON structure in the payload column: ```json { "event_type": "llm_end", "timestamp": "2025-09-18T10:30:00Z", "execution": { "run_id": "uuid-here", "parent_run_id": "", "custom_id": "user-123" }, "data": { "prompts": ["..."], "llm_type": "openai-chat", // LangChain's native LLM type "response": {"content": "..."}, "usage": {"total_tokens": 100} }, "raw": { // Complete dump of all callback arguments // Includes all kwargs plus positional args (serialized when possible) "response": {"generations": [...], "llm_output": {...}}, "run_id": "uuid-here", "parent_run_id": "", // ... all other arguments passed to the callback } } ``` ## Installation Options ```bash # Basic pip install langchain-callback-parquet-logger # With S3 support pip install "langchain-callback-parquet-logger[s3]" # With background retrieval support (OpenAI) pip install "langchain-callback-parquet-logger[background]" # Everything pip install "langchain-callback-parquet-logger[s3,background]" ``` ## License MIT ## Contributing Pull requests welcome! Keep it simple. ## Support [GitHub Issues](https://github.com/turbo3136/langchain-callback-parquet-logger/issues)
text/markdown
null
turbo3136 <turbo3136@gmail.com>
null
null
MIT
langchain, logging, parquet, llm, callback, monitoring
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Logging" ]
[]
null
null
>=3.10
[]
[]
[]
[ "pyarrow>=10.0.0", "langchain-core>=1.0.0", "pandas>=1.3.0", "pydantic>=2.0.0", "pytest>=7.0.0; extra == \"test\"", "pytest-asyncio>=0.21.0; extra == \"test\"", "pytest-mock>=3.10.0; extra == \"test\"", "pandas>=1.3.0; extra == \"test\"", "openai>=1.0.0; extra == \"background\"", "pandas>=1.3.0; extra == \"background\"", "tqdm>=4.60.0; extra == \"background\"", "boto3>=1.26.0; extra == \"s3\"" ]
[]
[]
[]
[ "Homepage, https://github.com/turbo3136/langchain-callback-parquet-logger", "Repository, https://github.com/turbo3136/langchain-callback-parquet-logger", "Issues, https://github.com/turbo3136/langchain-callback-parquet-logger/issues" ]
twine/6.2.0 CPython/3.11.14
2026-02-20T21:58:37.940313
langchain_callback_parquet_logger-3.2.1.tar.gz
47,225
f5/6a/05846c17c5b508038d85397574cc269f453bb5152039beb5786a37a0ee23/langchain_callback_parquet_logger-3.2.1.tar.gz
source
sdist
null
false
1a2fe6e0ac7f3d352303cb7b8879ab04
8702c00514f820beffe1c0366983337a52de6dd0460596452445ee95c866eb00
f56a05846c17c5b508038d85397574cc269f453bb5152039beb5786a37a0ee23
null
[ "LICENSE" ]
217
2.4
pytickersymbols
1.17.9
pytickersymbols provides access to google and yahoo ticker symbols
![Release Build](https://github.com/portfolioplus/pytickersymbols/workflows/Release%20Build/badge.svg) ![PyPI - Downloads](https://img.shields.io/pypi/dm/pytickersymbols?style=plastic) [![Coverage Status](https://coveralls.io/repos/github/portfolioplus/pytickersymbols/badge.svg?branch=master)](https://coveralls.io/github/portfolioplus/pytickersymbols?branch=master) [![Codacy Badge](https://app.codacy.com/project/badge/Grade/1385a87f773d47bc84336275a0182619)](https://www.codacy.com/gh/portfolioplus/pytickersymbols/dashboard?utm_source=github.com&amp;utm_medium=referral&amp;utm_content=portfolioplus/pytickersymbols&amp;utm_campaign=Badge_Grade) # pytickersymbols pytickersymbols provides access to google and yahoo ticker symbols for all stocks of the following indices: - [x] AEX - [x] BEL 20 - [x] CAC 40 - [x] CAC MID 60 - [x] DAX - [x] DOW JONES - [x] EURO STOXX 50 - [x] FTSE 100 - [x] IBEX 35 - [x] MDAX - [x] NASDAQ 100 - [x] OMX Helsinki 25 - [x] OMX Stockholm 30 - [x] S&P 100 - [x] S&P 500 - [x] S&P 600 - [x] SDAX - [x] Switzerland 20 - [x] TECDAX ## install ```shell pip3 install pytickersymbols ``` ## quick start Get all countries, indices and industries as follows: ```python from pytickersymbols import PyTickerSymbols stock_data = PyTickerSymbols() countries = stock_data.get_all_countries() indices = stock_data.get_all_indices() industries = stock_data.get_all_industries() ``` You can select all stocks of an index as follows: ```python from pytickersymbols import PyTickerSymbols stock_data = PyTickerSymbols() german_stocks = stock_data.get_stocks_by_index('DAX') uk_stocks = stock_data.get_stocks_by_index('FTSE 100') print(list(uk_stocks)) ``` If you are only interested in ticker symbols, then you should have a look at the following lines: ```python from pytickersymbols import PyTickerSymbols stock_data = PyTickerSymbols() # the naming conversation is get_{index_name}_{exchange_city}_{yahoo or google}_tickers dax_google = stock_data.get_dax_frankfurt_google_tickers() dax_yahoo = stock_data.get_dax_frankfurt_yahoo_tickers() sp100_yahoo = stock_data.get_sp_100_nyc_yahoo_tickers() sp500_google = stock_data.get_sp_500_nyc_google_tickers() dow_yahoo = stock_data.get_dow_jones_nyc_yahoo_tickers() # there are too many combination. Here is a complete list of all getters all_ticker_getter_names = list(filter( lambda x: ( x.endswith('_google_tickers') or x.endswith('_yahoo_tickers') ), dir(stock_data), )) print(all_ticker_getter_names) ``` ### Iterator Examples Use iterator-based APIs to stream results without building large lists: ```python from pytickersymbols import PyTickerSymbols stock_data = PyTickerSymbols() # Stream indices for index in stock_data.iter_all_indices(): print(index) # Stream all unique stocks for company in stock_data.iter_all_stocks(): print(company['name']) # Stream industries and countries for industry in stock_data.iter_all_industries(): pass # handle industry for country in stock_data.iter_all_countries(): pass # handle country # Stream Yahoo tickers for an index (flattened) for tickers in stock_data.iter_yahoo_ticker_symbols_by_index('DAX'): for ticker in tickers: print(ticker) # Stream tickers using exchange-specific dynamic methods for ticker in stock_data._iter_tickers_by_index('DAX', ('FRA:',), 'yahoo'): print(ticker) ``` ## Development ### Setting up the development environment This project uses Poetry for dependency management. To set up your development environment: ```shell # Install Poetry if you haven't already curl -sSL https://install.python-poetry.org | python3 - # Clone the repository git clone https://github.com/portfolioplus/pytickersymbols.git cd pytickersymbols # Install dependencies (including dev dependencies) poetry install # Activate the virtual environment poetry shell ``` ### Development Tools The `tools/` directory contains scripts for managing stock index data: - **build_indices.py**: End-to-end pipeline to parse Wikipedia, enrich with `stocks.yaml`, canonicalize names, and generate the Python data module. - **wiki_table_parser.py**: Utilities to parse Wikipedia tables configured via `index_sources.yaml`. - **enrich_indices.py**: Merge raw parsed data with `stocks.yaml` historical metadata and symbols. - **canonicalize_names.py**: Normalize company display names across indices using ISIN/Wikipedia. - **sync_canonical_to_stocks.py**: Propagate canonical names back to `stocks.yaml`. - **enrich_with_yfinance.py**: Optional enrichment helpers using Yahoo Finance. See [tools/README.md](tools/README.md) for detailed usage instructions. ### Running Tests ```shell poetry run pytest ``` ### Adding a New Index To add a new stock index to the library: 1. **Add the index to configuration** Edit [tools/index_sources.yaml](tools/index_sources.yaml) and add a new entry: ```yaml - name: NIKKEI 225 source: type: wikipedia url: https://en.wikipedia.org/wiki/Nikkei_225 table_title_regex: "Components" extract_company_info: true language_fallbacks: ["ja", "en"] columns: name: ["Company", "Name"] symbol: ["Ticker", "Symbol"] isin: ["ISIN"] sector: ["Sector", "Industry"] symbol_converter: - pattern: "^(.+)$" format: "{1}.T" # Add .T suffix for Tokyo Stock Exchange match: by: symbol ``` **Configuration options:** - `name`: Display name (must match stocks.yaml if merging with historical data) - `url`: Wikipedia page URL containing the index constituents table - `table_title_regex`: (Optional) Regex to match the table title - `extract_company_info`: Set to `true` to fetch additional details from company Wikipedia pages - `language_fallbacks`: List of Wikipedia language codes to try for ISIN lookup - `symbol_converter`: Rules to convert Wikipedia symbols to Yahoo Finance format - `columns`: Map Wikipedia table headers to data fields - `match.by`: Field to use for matching with historical data (`symbol`, `isin`, or `name`) 2. **Run the build pipeline** This will parse Wikipedia, enrich the data, and generate the Python module: ```shell cd tools python build_indices.py ``` This creates: - `indices_raw/<index_name>.json` - Raw parsed data from Wikipedia - `indices/<index_name>.yaml` - Enriched data merged with historical records - Updates `src/pytickersymbols/indices_data.py` - Generated Python module 3. **Test the new index** Verify the index is accessible: ```python from pytickersymbols import PyTickerSymbols stock_data = PyTickerSymbols() nikkei_stocks = stock_data.get_stocks_by_index('NIKKEI 225') print(list(nikkei_stocks)) ``` 4. **Run tests** Ensure everything works: ```shell poetry run pytest ``` 5. **Update the index list** Add a checkbox entry to the supported indices list at the top of this README. **Note:** The build pipeline automatically runs weekly via GitHub Actions to keep index data up to date. ### Performance Tips - Lookups for stocks by symbol and aggregated lists/sets are internally cached for speed. After calling `load_json()` or `load_yaml()`, caches are automatically rebuilt. - For large results, prefer generator variants to avoid materializing lists: `iter_all_indices()`, `iter_all_stocks()`, `iter_all_countries()`, `iter_all_industries()`, and `iter_yahoo_ticker_symbols_by_index()` / `iter_google_ticker_symbols_by_index()`. - Dynamic ticker getters follow the naming convention `get_{index_name}_{exchange_city}_{yahoo|google}_tickers` and return lists. Use `_iter_tickers_by_index()` for streaming if needed. ## issue tracker https://github.com/portfolioplus/pytickersymbols/issues
text/markdown
SlashGordon
slash.gordon.dev@gmail.com
null
null
MIT
null
[ "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14" ]
[]
null
null
<4.0,>=3.11
[]
[]
[]
[ "beautifulsoup4<5.0.0,>=4.12.0", "requests<3.0.0,>=2.31.0" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:57:16.069859
pytickersymbols-1.17.9.tar.gz
401,433
7c/c3/8df83a45ca3c1b4dfa72558f4b68b74e37006cbed3642fd4fc24e518dc18/pytickersymbols-1.17.9.tar.gz
source
sdist
null
false
fc02a2fde75238d1181f19e3dbbf46c8
d84ad92018bcebe1e419b104b5c7b209ee7faf16973f59ee4a5434a19667e7b9
7cc38df83a45ca3c1b4dfa72558f4b68b74e37006cbed3642fd4fc24e518dc18
null
[ "LICENSE" ]
293
2.4
opentelemetry-instrumentation-google-genai
0.7b0
OpenTelemetry
OpenTelemetry Google GenAI SDK Instrumentation ============================================== |pypi| .. |pypi| image:: https://badge.fury.io/py/opentelemetry-instrumentation-google-genai.svg :target: https://pypi.org/project/opentelemetry-instrumentation-google-genai/ This library adds instrumentation to the `Google GenAI SDK library <https://pypi.org/project/google-genai/>`_ to emit telemetry data following `Semantic Conventions for GenAI systems <https://opentelemetry.io/docs/specs/semconv/gen-ai/>`_. It adds trace spans for GenAI operations, events/logs for recording prompts/responses, and emits metrics that describe the GenAI operations in aggregate. Experimental ------------ This package is still experimental. The instrumentation may not be complete or correct just yet. Please see "TODOS.md" for a list of known defects/TODOs that are blockers to package stability. Installation ------------ If your application is already instrumented with OpenTelemetry, add this package to your requirements. :: pip install opentelemetry-instrumentation-google-genai If you don't have a Google GenAI SDK application, yet, try our `examples <examples>`_. Check out `zero-code example <examples/zero-code>`_ for a quick start. Usage ----- This section describes how to set up Google GenAI SDK instrumentation if you're setting OpenTelemetry up manually. Check out the `manual example <examples/manual>`_ for more details. Instrumenting all clients ************************* When using the instrumentor, all clients will automatically trace GenAI ``generate_content`` operations. You can also optionally capture prompts and responses as log events. Make sure to configure OpenTelemetry tracing, logging, metrics, and events to capture all telemetry emitted by the instrumentation. .. code-block:: python from opentelemetry.instrumentation.google_genai import GoogleGenAiSdkInstrumentor from google.genai import Client GoogleGenAiSdkInstrumentor().instrument() client = Client() response = client.models.generate_content( model="gemini-1.5-flash-002", contents="Write a short poem on OpenTelemetry." ) Limitations *********** When using the Google GenAI SDK with automatic function calling enabled, the OpenTelemetry instrumentation creates a span only for the top-level ``generate_content`` call. Internal model or tool calls triggered automatically by the SDK are executed within the SDK internals and are not traced as separate spans. Enabling message content ************************* Message content such as the contents of the prompt and response are not captured by default. To capture message content as log events, set the environment variable ``OTEL_INSTRUMENTATION_GENAI_CAPTURE_MESSAGE_CONTENT`` to ``true``. Uninstrument ************ To uninstrument clients, call the uninstrument method: .. code-block:: python from opentelemetry.instrumentation.google_genai import GoogleGenAiSdkInstrumentor GoogleGenAiSdkInstrumentor().instrument() # ... # Uninstrument all clients GoogleGenAiSdkInstrumentor().uninstrument() References ---------- * `Google Gen AI SDK Documentation <https://ai.google.dev/gemini-api/docs/sdks>`_ * `Google Gen AI SDK on GitHub <https://github.com/googleapis/python-genai>`_ * `Using Vertex AI with Google Gen AI SDK <https://cloud.google.com/vertex-ai/generative-ai/docs/sdks/overview>`_ * `OpenTelemetry Project <https://opentelemetry.io/>`_ * `OpenTelemetry Python Examples <https://github.com/open-telemetry/opentelemetry-python/tree/main/docs/examples>`_
text/x-rst
null
OpenTelemetry Authors <cncf-opentelemetry-contributors@lists.cncf.io>
null
null
null
null
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14" ]
[]
null
null
>=3.9
[]
[]
[]
[ "opentelemetry-api~=1.37", "opentelemetry-instrumentation<2,>=0.58b0", "opentelemetry-semantic-conventions<2,>=0.58b0", "opentelemetry-util-genai<0.4b0,>=0.3b0", "google-genai>=1.32.0; extra == \"instruments\"" ]
[]
[]
[]
[ "Homepage, https://github.com/open-telemetry/opentelemetry-python-contrib/tree/main/instrumentation-genai/opentelemetry-instrumentation-google-genai", "Repository, https://github.com/open-telemetry/opentelemetry-python-contrib" ]
twine/6.2.0 CPython/3.9.25
2026-02-20T21:56:23.116489
opentelemetry_instrumentation_google_genai-0.7b0.tar.gz
52,057
3f/59/595e5ed05715c47cf49ac7e30d0a4cf6ed41e00524401c46c9d92b84623c/opentelemetry_instrumentation_google_genai-0.7b0.tar.gz
source
sdist
null
false
90b19e2bf3f0385dc71a4550bbc28638
35158682dfd00201ef3864b47dda95d6062188e8a599ecabd383688e7eba2824
3f59595e5ed05715c47cf49ac7e30d0a4cf6ed41e00524401c46c9d92b84623c
Apache-2.0
[ "LICENSE" ]
2,655
2.4
motia
1.0.0rc22
Motia framework for III Engine
# Motia Framework for Python High-level framework for building workflows with the III Engine. ## Installation ```bash uv pip install motia ``` ## Usage ### Defining a Step ```python from motia import FlowContext, queue config = { "name": "process-data", "triggers": [queue("data.created")], "enqueues": ["data.processed"], } async def handler(data: dict, ctx: FlowContext) -> None: ctx.logger.info("Processing data", data) await ctx.enqueue({"topic": "data.processed", "data": data}) ``` ### API Steps ```python from motia import ApiRequest, ApiResponse, FlowContext, http config = { "name": "create-item", "triggers": [http("POST", "/items")], "emits": ["item.created"], } async def handler(req: ApiRequest, ctx: FlowContext) -> ApiResponse: ctx.logger.info("Creating item", req.body) await ctx.enqueue({"topic": "item.created", "data": req.body}) return ApiResponse(status=201, body={"id": "123"}) ``` ### Streams ```python from motia import Stream # Define a stream todo_stream = Stream[dict]("todos") # Use the stream item = await todo_stream.get("group-1", "item-1") await todo_stream.set("group-1", "item-1", {"title": "Buy milk"}) await todo_stream.delete("group-1", "item-1") items = await todo_stream.get_group("group-1") ``` ### Build & Publish ```bash python -m build uv publish --index cloudsmith dist/* ``` ## Features - Event-driven step definitions - API route handlers - Cron job support - Stream-based state management - Type-safe context with logging ## Testing ### Running Integration Tests Integration tests require a running III Engine instance. Make sure to have it built or installed before running tests. 1. Install dev dependencies: ```bash cd motia && uv sync --all-extras ``` ``` 3. Run tests: ```bash uv run pytest ``` ### Test Configuration Tests use non-default ports to avoid conflicts: - Engine WebSocket: `ws://localhost:49199` - HTTP API: `http://localhost:3199` Set `III_ENGINE_PATH` environment variable to point to the III engine binary. ### Test Coverage The integration test suite covers: - Bridge connection and function registration - API triggers (HTTP endpoints) - KV Server operations - PubSub messaging - Logging module - Motia framework integration - Stream operations (when available) - State management (when available)
text/markdown
III
null
null
null
null
null
[ "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12" ]
[]
null
null
>=3.10
[]
[]
[]
[ "iii-sdk==0.2.0", "pydantic>=2.0", "httpx>=0.27; extra == \"dev\"", "mypy>=1.8; extra == \"dev\"", "opentelemetry-api>=1.20; extra == \"dev\"", "opentelemetry-exporter-otlp>=1.20; extra == \"dev\"", "opentelemetry-sdk>=1.20; extra == \"dev\"", "pytest-asyncio>=0.23; extra == \"dev\"", "pytest>=8.0; extra == \"dev\"", "ruff>=0.2; extra == \"dev\"", "opentelemetry-api>=1.20; extra == \"otel\"", "opentelemetry-exporter-otlp>=1.20; extra == \"otel\"", "opentelemetry-sdk>=1.20; extra == \"otel\"" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:55:43.114335
motia-1.0.0rc22.tar.gz
98,591
59/21/e18b90392c6638de9aebbb2b24fa7c536d4c87e5fb9e4343b445338f378d/motia-1.0.0rc22.tar.gz
source
sdist
null
false
4c74ef6f87d11be108bae9a0a1a315fb
23c0832302c6df1fe7af7ce573fef17534b9b5fb583c93157a11380dda6d5c30
5921e18b90392c6638de9aebbb2b24fa7c536d4c87e5fb9e4343b445338f378d
null
[]
205
2.3
sport-activities-features
0.5.4
A minimalistic toolbox for extracting features from sport activity files
<p align="center"> <img width="200" src="https://raw.githubusercontent.com/firefly-cpp/sport-activities-features/main/.github/logo/sport_activities.png"> </p> <h1 align="center"> sport-activities-features --- A minimalistic toolbox for extracting features from sports activity files written in Python </h1> <p align="center"> <img alt="PyPI Version" src="https://img.shields.io/pypi/v/sport-activities-features.svg" href="https://pypi.python.org/pypi/sport-activities-features"> <img alt="PyPI - Python Version" src="https://img.shields.io/pypi/pyversions/sport-activities-features.svg"> <img alt="PyPI - Downloads" src="https://img.shields.io/pypi/dm/sport-activities-features.svg"> <img alt="Fedora package" src="https://img.shields.io/fedora/v/python3-sport-activities-features?color=blue&label=Fedora%20Linux&logo=fedora" href="https://src.fedoraproject.org/rpms/python-sport-activities-features"> <img alt="AUR package" src="https://img.shields.io/aur/version/python-sport-activities-features?color=blue&label=Arch%20Linux&logo=arch-linux" href="https://aur.archlinux.org/packages/python-sport-activities-features"> <img alt="Packaging status" src="https://repology.org/badge/tiny-repos/python:sport-activities-features.svg" href="https://repology.org/project/python:sport-activities-features/versions"> <img alt="Downloads" src="https://pepy.tech/badge/sport-activities-features" href="https://pepy.tech/project/sport-activities-features"> <img alt="GitHub license" src="https://img.shields.io/github/license/firefly-cpp/sport-activities-features.svg" href="https://github.com/firefly-cpp/sport-activities-features/blob/master/LICENSE"> <img alt="Documentation Status" src="https://readthedocs.org/projects/sport-activities-features/badge/?version=latest" href="https://sport-activities-features.readthedocs.io/en/latest/?badge=latest"> </p> <p align="center"> <img alt="GitHub repo size" src="https://img.shields.io/github/repo-size/firefly-cpp/sport-activities-features"> <img alt="GitHub commit activity" src="https://img.shields.io/github/commit-activity/w/firefly-cpp/sport-activities-features.svg"> <img alt="Average time to resolve an issue" src="http://isitmaintained.com/badge/resolution/firefly-cpp/sport-activities-features.svg" href="http://isitmaintained.com/project/firefly-cpp/sport-activities-features"> <img alt="Percentage of issues still open" src="http://isitmaintained.com/badge/open/firefly-cpp/sport-activities-features.svg" href="http://isitmaintained.com/project/firefly-cpp/sport-activities-features"> <img alt="All Contributors" src="https://img.shields.io/badge/all_contributors-6-orange.svg" href="#-contributors"> </p> <p align="center"> <img alt="DOI" src="https://img.shields.io/badge/DOI-10.1109/INES52918.2021.9512927-blue" href="https://doi.org/10.1109/INES52918.2021.9512927"> </p> <p align="center"> <a href="#-detailed-insights">🔍 Detailed insights</a> • <a href="#-installation">📦 Installation</a> • <a href="#-api">📮 API</a> • <a href="#-graphical-user-interface">💻 Graphical User Interface</a> • <a href="#️-historical-weather-data">🌦️ Historical weather data</a> • <a href="#-overpass-api-open-elevation-api--opentopodata-integration">🧩 Overpass API, Open Elevation API & OpenTopoData integration</a> <a href="#-examples">🚀 Examples</a> • <a href="#-license">🔑 License</a> • <a href="#-cite-us">📄 Cite us</a> • <a href="#-further-read">📖 Further read</a> • <a href="#-related-frameworks">🔗 Related frameworks</a> • <a href="#-contributors">🫂 Contributors</a> </p> ## Unleashing the Power of Sports Activity Analysis: A Framework Beyond Ordinary Metrics 🚀 Prepare to dive into the thrilling world of sports activity analysis, where hidden geographic, topological, and personalized data await their grand unveiling. In this captivating journey, we embark on a quest to extract the deepest insights from the wealth of information generated by monitoring sports activities. Brace yourself for a framework that transcends the limitations of conventional analysis techniques. 💪🔍 Traditional approaches often rely on integral metrics like total duration, total distance, and average heart rate, but they fall victim to the dreaded "overall metrics problem." These metrics fail to capture the essence of sports activities, omitting crucial components and leading to potentially flawed and misleading conclusions. They lack the ability to recognize distinct stages and phases of the activity, such as the invigorating warm-up, the endurance-testing main event, and the heart-pounding intervals. ⏱️🚴‍♀️📈 Fortunately, our sport-activities-framework rises above these limitations, revealing a comprehensive panorama of your sports activity files. This framework combines the power of identification and extraction methods to unlock a treasure trove of valuable data. Picture this :camera: : effortlessly identifying the number of hills, extracting average altitudes of these remarkable formations, measuring the total distance conquered on those inclines, and even deriving climbing ratios for a true measure of accomplishment (total distance of hills vs. total distance). But that's just the tip of the iceberg! The framework seamlessly integrates a multitude of extensions, including historical weather parsing, statistical evaluations, and ex-post visualizations that bring your data to life. 🗻📊🌦️ For those seeking to venture further, we invite you to explore the realms of scientific papers on data mining that delve into these captivating topics. Discover how our framework complements the world of generating and predicting automated sport training sessions, creating a harmonious synergy between theory and practice. 📚🔬💡 * **Free software:** MIT license * **Python versions:** 3.8.x, 3.9.x, 3.10.x, 3.11.x, 3.12.x * **Documentation:** https://sport-activities-features.readthedocs.io/en/latest * **Tested OS:** Windows, Ubuntu, Debian, Fedora, Alpine, Arch, macOS. **However, that does not mean it does not work on others.** ## 🔍 Detailed insights Prepare to be astounded by the capabilities of the sport-activities-features framework. It effortlessly handles TCX & GPX activity files and harnesses the power of the [Overpass API](https://wiki.openstreetmap.org/wiki/Overpass_API) nodes. Presenting the range of functions at your disposal: - **Unleash the integral metrics**: From total distance to total duration and even calorie count, witness the extraction of these vital statistics with a single glance. 📏⏰🔥 ([See example](examples/integral_metrics_extraction.py)) - **Conquer the peaks**: Ascend to new heights by extracting topographic features like the number of hills, their average altitudes, the total distance covered on these majestic slopes, and the thrilling climbing ratio. Prepare for a breathtaking adventure! ⛰️📈🧗‍♂️ ([See example](examples/hill_data_extraction.py)) - **Embark on a visual journey**: Immerse yourself in the beauty of your accomplishments as you plot the identified hills on a mesmerizing map. Witness the landscape come alive before your eyes. 🗺️🏞️🖌️ ([See example](examples/draw_map_with_identified_hills.py)) - **Embrace the rhythm of intervals**: Explore the intervals within your sports activities, uncovering their numbers, durations, distances, and heart rates. Unveil the heartbeat of your performance! 🏃‍♀️📊💓 ([See example](examples/draw_map_with_identified_intervals.py)) - **Calculate the training loads**: Dive deep into the intricate world of training loads and discover the Banister TRIMP and Lucia TRIMP methods. Gain invaluable insights into optimizing your training regimen. 📈⚖️🏋️‍♂️ ([See example](examples/calculate_training_load.py)) - **Weather the storm**: Unlock the power of historical weather data from external services, adding a fascinating layer of context to your sports activities. ☀️🌧️⛈️ - **Unveil the secrets within coordinates**: Explore the integral metrics of your activities within specific geographical areas, uncovering valuable data on distance, heart rate, and speed. Peer into the depths of your performance! 🌍📍📉 ([See example](examples/extract_data_inside_area.py)) - **Embrace randomness**: Extract activities from CSV files and indulge in the excitement of randomly selecting a specific number of activities. Embrace the element of surprise! 🎲📂🎉 ([See example](examples/extract_random_activities_from_csv.py)) - **Conquer the dead ends**: Unravel the mysteries of your sports activities by identifying the dead ends. Prepare to navigate the uncharted territories of your performance! 🚧🗺️🔍 ([See example](examples/dead_end_extraction.py)) - **Unlock the format**: Seamlessly convert TCX files to GPX, opening doors to even more possibilities. Adapt and conquer! ⚙️🔄✨ ([See example](examples/convert_tcx_to_gpx.py)) And that's just the beginning! The sport-activities-framework holds countless other features, awaiting your exploration. Brace yourself for an exhilarating journey of discovery, where the ordinary becomes extraordinary, and your sports activities come alive like never before. 🌟🔥🏃‍♂️ The framework comes with two (testing) [benchmark datasets](https://github.com/firefly-cpp/sports-activity-dataset-collections), which are freely available to download from: [DATASET1](http://iztok-jr-fister.eu/static/publications/Sport5.zip), [DATASET2](http://iztok-jr-fister.eu/static/css/datasets/Sport.zip). ## 📦 Installation ### pip Install sport-activities-features with pip: ```sh pip install sport-activities-features ``` ### Alpine Linux To install sport-activities-features on Alpine, use: ```sh $ apk add py3-sport-activities-features ``` ### Fedora Linux To install sport-activities-features on Fedora, use: ```sh $ dnf install python3-sport-activities-features ``` ### Arch Linux To install sport-activities-features on Arch Linux, please use an [AUR helper](https://wiki.archlinux.org/title/AUR_helpers): ```sh $ yay -Syyu python-sport-activities-features ``` ## 📮 API There is a simple API for remote work with the sport-activities-features package available [here](https://github.com/alenrajsp/sport-activities-features-api). ## 💻 Graphical User Interface There is a simple Graphical User Interface for the sport-activities-features package available [here](https://github.com/firefly-cpp/sport-activities-features-gui). ## 🌦️ Historical weather data Weather data parsed is collected from the [Visual Crossing Weather API](https://www.visualcrossing.com/). Please note that this is an external unaffiliated service, and users must register to use the API. The service has a free tier (1000 Weather reports/day) but is otherwise operating on a pay-as-you-go model. For pricing and terms of use, please read the [official documentation](https://www.visualcrossing.com/weather-data-editions) of the API provider. ## 🧩 Overpass API, Open Elevation API & OpenTopoData integration Without performing activities, we can use the [OpenStreetMap](https://www.openstreetmap.org/) for the identification of hills, total ascent, and descent. This is done using the [Overpass API](https://wiki.openstreetmap.org/wiki/Overpass_API) which is a read-only API that allows querying of OSM map data. In addition to that altitude, data is retrieved by using the [Open-Elevation API](https://open-elevation.com/) or [OpenTopoData API](https://www.opentopodata.org/) which are open-source and free alternatives to the Google Elevation API. Both of the solutions can be used by using free publicly acessible APIs ([Overpass](https://wiki.openstreetmap.org/wiki/Overpass_API), [Open-Elevation](https://open-elevation.com/#public-api), [OpenTopoData](https://www.opentopodata.org/#public-api)) or can be self hosted on a server or as a Docker container ([Overpass](https://wiki.openstreetmap.org/wiki/Overpass_API/Installation), [Open-Elevation](https://github.com/Jorl17/open-elevation/blob/master/docs/host-your-own.md), [OpenTopoData](https://www.opentopodata.org/server/)). ## 🚀 Examples ### Reading files #### (*.TCX) ```python from sport_activities_features.tcx_manipulation import TCXFile # Class for reading TCX files tcx_file=TCXFile() tcx_exercise = tcx_file.read_one_file("path_to_the_file") data = tcx_file.extract_activity_data(tcx_exercise) # Represents data as dictionary of lists # Alternative choice data = tcx_file.extract_activity_data(tcx_exercise, numpy_array= True) # Represents data as dictionary of numpy.arrays ``` #### (*.GPX) ```python from sport_activities_features.gpx_manipulation import GPXFile # Class for reading GPX files gpx_file=GPXFile() # Read the file and generate a dictionary with gpx_exercise = gpx_file.read_one_file("path_to_the_file") data = gpx_file.extract_activity_data(gpx_exercise) # Represents data as dictionary of lists # Alternative choice data = gpx_file.extract_activity_data(gpx_exercise, numpy_array= True) # Represents data as dictionary of numpy.arrays ``` ### Extraction of topographic features ```python from sport_activities_features.hill_identification import HillIdentification from sport_activities_features.tcx_manipulation import TCXFile from sport_activities_features.topographic_features import TopographicFeatures from sport_activities_features.plot_data import PlotData # Read TCX file tcx_file = TCXFile() tcx_exercise = tcx_file.read_one_file("path_to_the_file") activity = tcx_file.extract_activity_data(tcx_exercise) # Detect hills in data Hill = HillIdentification(activity['altitudes'], 30) Hill.identify_hills() all_hills = Hill.return_hills() # Extract features from data Top = TopographicFeatures(all_hills) num_hills = Top.num_of_hills() avg_altitude = Top.avg_altitude_of_hills(activity['altitudes']) avg_ascent = Top.avg_ascent_of_hills(activity['altitudes']) distance_hills = Top.distance_of_hills(activity['positions']) hills_share = Top.share_of_hills(distance_hills, activity['total_distance']) ``` ### Extraction of intervals ```python import sys sys.path.append('../') from sport_activities_features.interval_identification import IntervalIdentificationByPower, IntervalIdentificationByHeartrate from sport_activities_features.tcx_manipulation import TCXFile # Reading the TCX file tcx_file = TCXFile() tcx_exercise = tcx_file.read_one_file("path_to_the_file") activity = tcx_file.extract_activity_data(tcx_exercise) # Identifying the intervals in the activity by power Intervals = IntervalIdentificationByPower(activity["distances"], activity["timestamps"], activity["altitudes"], 70) Intervals.identify_intervals() all_intervals = Intervals.return_intervals() # Identifying the intervals in the activity by heart rate Intervals = IntervalIdentificationByHeartrate(activity["timestamps"], activity["altitudes"], activity["heartrates"]) Intervals.identify_intervals() all_intervals = Intervals.return_intervals() ``` ### Parsing of Historical weather data from an external service ```python from sport_activities_features import WeatherIdentification from sport_activities_features import TCXFile # Read TCX file tcx_file = TCXFile() tcx_exercise = tcx_file.read_one_file("path_to_the_file") tcx_data = tcx_file.extract_activity_data(tcx_exercise) # Configure visual crossing api key visual_crossing_api_key = "weather_api_key" # https://www.visualcrossing.com/weather-api # Explanation of elements - https://www.visualcrossing.com/resources/documentation/weather-data/weather-data-documentation/ weather = WeatherIdentification(tcx_data['positions'], tcx_data['timestamps'], visual_crossing_api_key) weatherlist = weather.get_weather(time_delta=30) tcx_weather = weather.get_average_weather_data(timestamps=tcx_data['timestamps'],weather=weatherlist) # Add weather to TCX data tcx_data.update({'weather':tcx_weather}) ``` The weather list is of the following type: ```json [ { "temperature": 14.3, "maximum_temperature": 14.3, "minimum_temperature": 14.3, "wind_chill": null, "heat_index": null, "solar_radiation": null, "precipitation": 0.0, "sea_level_pressure": 1021.6, "snow_depth": null, "wind_speed": 6.9, "wind_direction": 129.0, "wind_gust": null, "visibility": 40.0, "cloud_cover": 54.3, "relative_humidity": 47.6, "dew_point": 3.3, "weather_type": "", "conditions": "Partially cloudy", "date": "2016-04-02T17:26:09+00:00", "location": [ 46.079871179535985, 14.738618675619364 ], "index": 0 }, ... ] ``` ### Extraction of integral metrics ```python import sys sys.path.append('../') from sport_activities_features.tcx_manipulation import TCXFile # Read TCX file tcx_file = TCXFile() tcx_exercise = tcx_file.read_one_file("path_to_the_file") integral_metrics = tcx_file.extract_integral_metrics(tcx_exercise) print(integral_metrics) ``` ### Weather data extraction ```python from sport_activities_features.weather_identification import WeatherIdentification from sport_activities_features.tcx_manipulation import TCXFile #read TCX file tcx_file = TCXFile() tcx_exercise = tcx_file.read_one_file("path_to_the_file") tcx_data = tcx_file.extract_activity_data(tcx_exercise) #configure visual crossing api key visual_crossing_api_key = "API_KEY" # https://www.visualcrossing.com/weather-api #return weather objects weather = WeatherIdentification(tcx_data['positions'], tcx_data['timestamps'], visual_crossing_api_key) weatherlist = weather.get_weather() ``` ### Using Overpass queried Open Street Map nodes ```python import overpy from sport_activities_features.overpy_node_manipulation import OverpyNodesReader # External service Overpass API (https://wiki.openstreetmap.org/wiki/Overpass_API) (can be self-hosted) overpass_api = "https://lz4.overpass-api.de/api/interpreter" # External service Open Elevation API (https://api.open-elevation.com/api/v1/lookup) (can be self-hosted) open_elevation_api = "https://api.open-elevation.com/api/v1/lookup" # OSM Way (https://wiki.openstreetmap.org/wiki/Way) open_street_map_way = 164477980 overpass_api = overpy.Overpass(url=overpass_api) # Get an example Overpass way q = f"""(way({open_street_map_way});<;);out geom;""" query = overpass_api.query(q) # Get nodes of an Overpass way nodes = query.ways[0].get_nodes(resolve_missing=True) # Extract basic data from nodes (you can, later on, use Hill Identification and Hill Data Extraction on them) overpy_reader = OverpyNodesReader(open_elevation_api=open_elevation_api) # Returns { # 'positions': positions, 'altitudes': altitudes, 'distances': distances, 'total_distance': total_distance # } data = overpy_reader.read_nodes(nodes) ``` ### Extraction of data inside the area ```python import numpy as np import sys sys.path.append('../') from sport_activities_features.area_identification import AreaIdentification from sport_activities_features.tcx_manipulation import TCXFile # Reading the TCX file. tcx_file = TCXFile() tcx_exercise = tcx_file.read_one_file("path_to_the_file") activity = tcx_file.extract_activity_data(tcx_exercise) # Converting the read data to arrays. positions = np.array([*activity['positions']]) distances = np.array([*activity['distances']]) timestamps = np.array([*activity['timestamps']]) heartrates = np.array([*activity['heartrates']]) # Area coordinates should be given in clockwise orientation in the form of 3D array (number_of_hulls, hull_coordinates, 2). # Holes in area are permitted. area_coordinates = np.array([[[10, 10], [10, 50], [50, 50], [50, 10]], [[19, 19], [19, 21], [21, 21], [21, 19]]]) # Extracting the data inside the given area. area = AreaIdentification(positions, distances, timestamps, heartrates, area_coordinates) area.identify_points_in_area() area_data = area.extract_data_in_area() ``` ### Identify interruptions ```python from sport_activities_features.interruptions.interruption_processor import InterruptionProcessor from sport_activities_features.tcx_manipulation import TCXFile """ Identify interruption events from a TCX or GPX file. """ # read TCX file (also works with GPX files) tcx_file = TCXFile() tcx_exercise = tcx_file.read_one_file("path_to_the_file") tcx_data = tcx_file.extract_activity_data(tcx_exercise) """ Time interval = time before and after the start of an event Min speed = Threshold speed to trigger an event/interruption (trigger when under min_speed) overpass_api_url = Set to something self-hosted, or use a public instance from https://wiki.openstreetmap.org/wiki/Overpass_API """ interruptionProcessor = InterruptionProcessor(time_interval=60, min_speed=2, overpass_api_url="url_to_overpass_api") """ If classify is set to true, also discover if interruptions are close to intersections. Returns a list of [ExerciseEvent] """ events = interruptionProcessor.events(tcx_data, True) ``` ### Overpy (Overpass API) node manipulation Generate TCXFile parsed like data object from overpy.Node objects ```python import overpy from sport_activities_features.overpy_node_manipulation import OverpyNodesReader # External service Overpass API (https://wiki.openstreetmap.org/wiki/Overpass_API) (can be self-hosted) overpass_api = "https://lz4.overpass-api.de/api/interpreter" # External service Open Elevation API (https://api.open-elevation.com/api/v1/lookup) (can be self-hosted) open_elevation_api = "https://api.open-elevation.com/api/v1/lookup" # OSM Way (https://wiki.openstreetmap.org/wiki/Way) open_street_map_way = 164477980 overpass_api = overpy.Overpass(url=overpass_api) # Get an example Overpass way q = f"""(way({open_street_map_way});<;);out geom;""" query = overpass_api.query(q) # Get nodes of an Overpass way nodes = query.ways[0].get_nodes(resolve_missing=True) # Extract basic data from nodes (you can, later on, use Hill Identification and Hill Data Extraction on them) overpy_reader = OverpyNodesReader(open_elevation_api=open_elevation_api) # Returns { # 'positions': positions, 'altitudes': altitudes, 'distances': distances, 'total_distance': total_distance # } data = overpy_reader.read_nodes(nodes) ``` ### Missing elevation data extraction ```python from sport_activities_features import ElevationIdentification from sport_activities_features import TCXFile tcx_file = TCXFile() tcx_exercise = tcx_file.read_one_file("path_to_the_file") tcx_data = tcx_file.extract_activity_data(tcx_exercise) elevations = ElevationIdentification(tcx_data['positions']) """Adds tcx_data['elevation'] = eg. [124, 21, 412] for each position""" tcx_data.update({'elevations':elevations}) ``` ### Example of a visualization of the area detection ![Area Figure](https://i.imgur.com/Iz8Ga3B.png) ### Example of visualization of dead-end identification ![Dead End Figure](https://imgur.com/LgZzCFS.png) ## 🔑 License This package is distributed under the MIT License. This license can be found online at <http://www.opensource.org/licenses/MIT>. ## Disclaimer This framework is provided as-is, and there are no guarantees that it fits your purposes or that it is bug-free. Use it at your own risk! ## 📄 Cite us I. Jr. Fister, L. Lukač, A. Rajšp, I. Fister, L. Pečnik and D. Fister, "[A minimalistic toolbox for extracting features from sport activity files](http://iztok-jr-fister.eu/static/publications/294.pdf)", 2021 IEEE 25th International Conference on Intelligent Engineering Systems (INES), 2021, pp. 121-126, doi: [10.1109/INES52918.2021.9512927](http://dx.doi.org/10.1109/INES52918.2021.9512927). ## 📖 Further read [1] [Awesome Computational Intelligence in Sports](https://github.com/firefly-cpp/awesome-computational-intelligence-in-sports) ## 🔗 Related frameworks [1] [AST-Monitor: A wearable Raspberry Pi computer for cyclists](https://github.com/firefly-cpp/AST-Monitor) [2] [TCXReader.jl: Julia package designed for parsing TCX files](https://github.com/firefly-cpp/TCXReader.jl) [3] [TCXWriter: A Tiny Library for writing/creating TCX files on Arduino](https://github.com/firefly-cpp/tcxwriter) ## 🫂 Contributors Thanks go to these wonderful people ([emoji key](https://allcontributors.org/docs/en/emoji-key)): <!-- ALL-CONTRIBUTORS-LIST:START - Do not remove or modify this section --> <!-- prettier-ignore-start --> <!-- markdownlint-disable --> <table> <tbody> <tr> <td align="center" valign="top" width="14.28%"><a href="https://github.com/alenrajsp"><img src="https://avatars.githubusercontent.com/u/27721714?v=4?s=100" width="100px;" alt="alenrajsp"/><br /><sub><b>alenrajsp</b></sub></a><br /><a href="https://github.com/firefly-cpp/sport-activities-features/commits?author=alenrajsp" title="Code">💻</a> <a href="https://github.com/firefly-cpp/sport-activities-features/commits?author=alenrajsp" title="Tests">⚠️</a> <a href="#example-alenrajsp" title="Examples">💡</a> <a href="https://github.com/firefly-cpp/sport-activities-features/commits?author=alenrajsp" title="Documentation">📖</a> <a href="#ideas-alenrajsp" title="Ideas, Planning, & Feedback">🤔</a> <a href="https://github.com/firefly-cpp/sport-activities-features/issues?q=author%3Aalenrajsp" title="Bug reports">🐛</a> <a href="#maintenance-alenrajsp" title="Maintenance">🚧</a></td> <td align="center" valign="top" width="14.28%"><a href="http://www.iztok-jr-fister.eu/"><img src="https://avatars.githubusercontent.com/u/1633361?v=4?s=100" width="100px;" alt="Iztok Fister Jr."/><br /><sub><b>Iztok Fister Jr.</b></sub></a><br /><a href="https://github.com/firefly-cpp/sport-activities-features/commits?author=firefly-cpp" title="Code">💻</a> <a href="https://github.com/firefly-cpp/sport-activities-features/issues?q=author%3Afirefly-cpp" title="Bug reports">🐛</a> <a href="https://github.com/firefly-cpp/sport-activities-features/commits?author=firefly-cpp" title="Tests">⚠️</a> <a href="#example-firefly-cpp" title="Examples">💡</a> <a href="https://github.com/firefly-cpp/sport-activities-features/commits?author=firefly-cpp" title="Documentation">📖</a> <a href="#ideas-firefly-cpp" title="Ideas, Planning, & Feedback">🤔</a> <a href="#mentoring-firefly-cpp" title="Mentoring">🧑‍🏫</a> <a href="#platform-firefly-cpp" title="Packaging/porting to new platform">📦</a> <a href="#maintenance-firefly-cpp" title="Maintenance">🚧</a> <a href="#data-firefly-cpp" title="Data">🔣</a></td> <td align="center" valign="top" width="14.28%"><a href="https://github.com/luckyLukac"><img src="https://avatars.githubusercontent.com/u/73126820?v=4?s=100" width="100px;" alt="luckyLukac"/><br /><sub><b>luckyLukac</b></sub></a><br /><a href="#ideas-luckyLukac" title="Ideas, Planning, & Feedback">🤔</a> <a href="https://github.com/firefly-cpp/sport-activities-features/commits?author=luckyLukac" title="Code">💻</a> <a href="https://github.com/firefly-cpp/sport-activities-features/issues?q=author%3AluckyLukac" title="Bug reports">🐛</a> <a href="https://github.com/firefly-cpp/sport-activities-features/commits?author=luckyLukac" title="Tests">⚠️</a> <a href="#example-luckyLukac" title="Examples">💡</a></td> <td align="center" valign="top" width="14.28%"><a href="https://github.com/rhododendrom"><img src="https://avatars.githubusercontent.com/u/3198785?v=4?s=100" width="100px;" alt="rhododendrom"/><br /><sub><b>rhododendrom</b></sub></a><br /><a href="https://github.com/firefly-cpp/sport-activities-features/commits?author=rhododendrom" title="Code">💻</a> <a href="#design-rhododendrom" title="Design">🎨</a> <a href="https://github.com/firefly-cpp/sport-activities-features/commits?author=rhododendrom" title="Documentation">📖</a> <a href="#ideas-rhododendrom" title="Ideas, Planning, & Feedback">🤔</a></td> <td align="center" valign="top" width="14.28%"><a href="https://github.com/lukapecnik"><img src="https://avatars.githubusercontent.com/u/23029992?v=4?s=100" width="100px;" alt="Luka Pečnik"/><br /><sub><b>Luka Pečnik</b></sub></a><br /><a href="https://github.com/firefly-cpp/sport-activities-features/commits?author=lukapecnik" title="Code">💻</a> <a href="https://github.com/firefly-cpp/sport-activities-features/commits?author=lukapecnik" title="Documentation">📖</a> <a href="https://github.com/firefly-cpp/sport-activities-features/commits?author=lukapecnik" title="Tests">⚠️</a> <a href="https://github.com/firefly-cpp/sport-activities-features/issues?q=author%3Alukapecnik" title="Bug reports">🐛</a></td> <td align="center" valign="top" width="14.28%"><a href="https://github.com/spelap"><img src="https://avatars.githubusercontent.com/u/19300429?v=4?s=100" width="100px;" alt="spelap"/><br /><sub><b>spelap</b></sub></a><br /><a href="https://github.com/firefly-cpp/sport-activities-features/commits?author=spelap" title="Code">💻</a></td> <td align="center" valign="top" width="14.28%"><a href="http://carlosal1015.github.io"><img src="https://avatars.githubusercontent.com/u/21283014?v=4?s=100" width="100px;" alt="Oromion"/><br /><sub><b>Oromion</b></sub></a><br /><a href="#maintenance-carlosal1015" title="Maintenance">🚧</a> <a href="https://github.com/firefly-cpp/sport-activities-features/issues?q=author%3Acarlosal1015" title="Bug reports">🐛</a></td> </tr> <tr> <td align="center" valign="top" width="14.28%"><a href="https://github.com/KoprivcLuka"><img src="https://avatars.githubusercontent.com/u/45632449?v=4?s=100" width="100px;" alt="Luka Koprivc"/><br /><sub><b>Luka Koprivc</b></sub></a><br /><a href="https://github.com/firefly-cpp/sport-activities-features/issues?q=author%3AKoprivcLuka" title="Bug reports">🐛</a></td> <td align="center" valign="top" width="14.28%"><a href="https://github.com/garyjellyarms"><img src="https://avatars.githubusercontent.com/u/79595804?v=4?s=100" width="100px;" alt="Nejc Graj"/><br /><sub><b>Nejc Graj</b></sub></a><br /><a href="https://github.com/firefly-cpp/sport-activities-features/issues?q=author%3Agaryjellyarms" title="Bug reports">🐛</a></td> <td align="center" valign="top" width="14.28%"><a href="https://github.com/Mtvrt123"><img src="https://avatars.githubusercontent.com/u/50879568?v=4?s=100" width="100px;" alt="MlinaricNejc"/><br /><sub><b>MlinaricNejc</b></sub></a><br /><a href="https://github.com/firefly-cpp/sport-activities-features/issues?q=author%3AMtvrt123" title="Bug reports">🐛</a></td> <td align="center" valign="top" width="14.28%"><a href="https://github.com/KukovecRok"><img src="https://avatars.githubusercontent.com/u/33880044?v=4?s=100" width="100px;" alt="Tatookie"/><br /><sub><b>Tatookie</b></sub></a><br /><a href="https://github.com/firefly-cpp/sport-activities-features/commits?author=KukovecRok" title="Code">💻</a> <a href="https://github.com/firefly-cpp/sport-activities-features/issues?q=author%3AKukovecRok" title="Bug reports">🐛</a> <a href="https://github.com/firefly-cpp/sport-activities-features/commits?author=KukovecRok" title="Tests">⚠️</a> <a href="#example-KukovecRok" title="Examples">💡</a> <a href="#maintenance-KukovecRok" title="Maintenance">🚧</a></td> <td align="center" valign="top" width="14.28%"><a href="https://github.com/zala-lahovnik"><img src="https://avatars.githubusercontent.com/u/105444201?v=4?s=100" width="100px;" alt="Zala Lahovnik"/><br /><sub><b>Zala Lahovnik</b></sub></a><br /><a href="https://github.com/firefly-cpp/sport-activities-features/commits?author=zala-lahovnik" title="Documentation">📖</a> <a href="https://github.com/firefly-cpp/sport-activities-features/commits?author=zala-lahovnik" title="Code">💻</a></td> <td align="center" valign="top" width="14.28%"><a href="https://github.com/lahovniktadej"><img src="https://avatars.githubusercontent.com/u/57890734?v=4?s=100" width="100px;" alt="Tadej Lahovnik"/><br /><sub><b>Tadej Lahovnik</b></sub></a><br /><a href="https://github.com/firefly-cpp/sport-activities-features/commits?author=lahovniktadej" title="Documentation">📖</a></td> <td align="center" valign="top" width="14.28%"><a href="https://github.com/HlisTilen"><img src="https://avatars.githubusercontent.com/u/44733158?v=4?s=100" width="100px;" alt="HlisTilen"/><br /><sub><b>HlisTilen</b></sub></a><br /><a href="https://github.com/firefly-cpp/sport-activities-features/commits?author=HlisTilen" title="Documentation">📖</a></td> </tr> <tr> <td align="center" valign="top" width="14.28%"><a href="https://github.com/MihaMi27"><img src="https://avatars.githubusercontent.com/u/82605811?v=4?s=100" width="100px;" alt="Miha Mirt"/><br /><sub><b>Miha Mirt</b></sub></a><br /><a href="https://github.com/firefly-cpp/sport-activities-features/commits?author=MihaMi27" title="Code">💻</a></td> </tr> </tbody> </table> <!-- markdownlint-restore --> <!-- prettier-ignore-end --> <!-- ALL-CONTRIBUTORS-LIST:END --> This project follows the [all-contributors](https://github.com/all-contributors/all-contributors) specification. Contributions of any kind are welcome!
text/markdown
Iztok Fister Jr.
iztok@iztok-jr-fister.eu
null
null
MIT
computational intelligence, cycling, data mining, datasets, gpx, optimization, sport activities, tcx
[ "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13" ]
[]
https://github.com/firefly-cpp/sport-activities-features
null
<4.0.0,>=3.9.0
[]
[]
[]
[ "matplotlib<4.0.0,>=3.3.3", "geopy<3.0.0,>=2.0.0", "overpy<0.7,>=0.6", "geotiler<0.16.0,>=0.15.1", "numpy>=1.26.4", "tcxreader<0.5.0,>=0.4.11", "pandas", "tcx2gpx==0.1.4", "gpxpy==1.4.2", "setuptools<70" ]
[]
[]
[]
[ "Homepage, https://github.com/firefly-cpp/sport-activities-features", "Repository, https://github.com/firefly-cpp/sport-activities-features", "Documentation, https://sport-activities-features.readthedocs.io/en/latest/" ]
poetry/2.1.4 CPython/3.14.2 Linux/6.18.7-200.fc43.x86_64
2026-02-20T21:55:41.819012
sport_activities_features-0.5.4.tar.gz
54,928
16/11/b280b5bb53aa00123f36bc0c302be2746c0e4fe64c7cab0796615551cb81/sport_activities_features-0.5.4.tar.gz
source
sdist
null
false
9d84af430e74221652f853319319da2b
c83e73c581427af92a0a67da0178386c91cfb93706f1a3f3e3cd41ee03435d80
1611b280b5bb53aa00123f36bc0c302be2746c0e4fe64c7cab0796615551cb81
null
[]
222
2.4
inspect-flow
0.4.1
Inspect Flow is a workflow stack built on Inspect AI that enables research organizations to run AI evaluations at scale
<img src="docs/images/icon-dark.svg" alt="Inspect Flow" width="50" height="50"> # Inspect Flow Workflow orchestration for [Inspect AI](https://inspect.aisi.org.uk/) that enables you to run evaluations at scale with repeatability and maintainability. ## Why Inspect Flow? As evaluation workflows grow in complexity—running multiple tasks across different models with varying parameters—managing these experiments becomes challenging. Inspect Flow addresses this by providing: 1. **Declarative Configuration**: Define complex evaluations with tasks, models, and parameters in type-safe schemas 2. **Repeatable & Shareable**: Encapsulated definitions of tasks, models, configurations, and Python dependencies ensure experiments can be reliably repeated and shared 3. **Powerful Defaults**: Define defaults once and reuse them everywhere with automatic inheritance 4. **Parameter Sweeping**: Matrix patterns for systematic exploration across tasks, models, and hyperparameters Inspect Flow is designed for researchers and engineers running systematic AI evaluations who need to scale beyond ad-hoc scripts. ## Getting Started ### Prerequisites Before using Inspect Flow, you should: - Have familiarity with [Inspect AI](https://inspect.aisi.org.uk/) - Have an existing Inspect evaluation or use one from [inspect-evals](https://github.com/UKGovernmentBEIS/inspect_evals) ### Installation ```bash pip install inspect-flow ``` ### Optional: VS Code extension Optionally install the [Inspect AI VS Code Extension](https://inspect.aisi.org.uk/vscode.html) which includes features for viewing evaluation log files. ## Basic Example `FlowSpec` is the main entrypoint for defining evaluation runs. At its core, it takes a list of tasks to run. Here's a simple example that runs two evaluations: ```python from inspect_flow import FlowSpec, FlowTask FlowSpec( log_dir="logs", tasks=[ FlowTask( name="inspect_evals/gpqa_diamond", model="openai/gpt-4o", ), FlowTask( name="inspect_evals/mmlu_0_shot", model="openai/gpt-4o", ), ], ) ``` To run the evaluations, run the following command in your shell. This will create a virtual environment for this spec run and install the dependencies. Note that task and model dependencies (like the `inspect-evals` and `openai` Python packages) are inferred and installed automatically. ```bash flow run config.py ``` This will run both tasks and display progress in your terminal. ![Progress bar in terminal](docs/images/config_progress_terminal.png) ### Python API You can run evaluations from Python instead of the command line. ```python from inspect_flow import FlowSpec, FlowTask from inspect_flow.api import run spec = FlowSpec( log_dir="logs", tasks=[ FlowTask( name="inspect_evals/gpqa_diamond", model="openai/gpt-4o", ), FlowTask( name="inspect_evals/mmlu_0_shot", model="openai/gpt-4o", ), ], ) run(spec=spec) ``` ## Matrix Functions Often you'll want to evaluate multiple tasks across multiple models. Rather than manually defining every combination, use `tasks_matrix` to generate all task-model pairs: ```python from inspect_flow import FlowSpec, tasks_matrix FlowSpec( log_dir="logs", tasks=tasks_matrix( task=[ "inspect_evals/gpqa_diamond", "inspect_evals/mmlu_0_shot", ], model=[ "openai/gpt-5", "openai/gpt-5-mini", ], ), ) ``` To preview the expanded config before running it, you can run the following command in your shell to ensure the generated config is the one that you intend to run. ```bash flow config matrix.py ``` This command outputs the expanded configuration showing all 4 task-model combinations (2 tasks × 2 models). ```yaml log_dir: logs dependencies: - inspect-evals tasks: - name: inspect_evals/gpqa_diamond model: name: openai/gpt-5 - name: inspect_evals/gpqa_diamond model: name: openai/gpt-5-mini - name: inspect_evals/mmlu_0_shot model: name: openai/gpt-5 - name: inspect_evals/mmlu_0_shot model: name: openai/gpt-5-mini ``` `tasks_matrix` and `models_matrix` are powerful functions that can operate on multiple levels of nested matrixes which enable sophisticated parameter sweeping. Let's say you want to explore different reasoning efforts across models—you can achieve this with the `models_matrix` function. ```python from inspect_ai.model import GenerateConfig from inspect_flow import FlowSpec, models_matrix, tasks_matrix FlowSpec( log_dir="logs", tasks=tasks_matrix( task=[ "inspect_evals/gpqa_diamond", "inspect_evals/mmmu_0_shot", ], model=models_matrix( model=[ "openai/gpt-5", "openai/gpt-5-mini", ], config=[ GenerateConfig(reasoning_effort="minimal"), GenerateConfig(reasoning_effort="low"), GenerateConfig(reasoning_effort="medium"), GenerateConfig(reasoning_effort="high"), ], ), ), ) ``` For even more concise parameter sweeping, use `configs_matrix` to generate configuration variants. This produces the same 16 evaluations (2 tasks × 2 models × 4 reasoning levels) as above, but with less boilerplate: ```python from inspect_flow import FlowSpec, configs_matrix, models_matrix, tasks_matrix FlowSpec( log_dir="logs", tasks=tasks_matrix( task=[ "inspect_evals/gpqa_diamond", "inspect_evals/mmmu_0_shot", ], model=models_matrix( model=[ "openai/gpt-5", "openai/gpt-5-mini", ], config=configs_matrix( reasoning_effort=["minimal", "low", "medium", "high"], ), ), ), ) ``` ### Run evaluations Before running evaluations, preview the resolved configuration with `--dry-run`: ```bash flow run matrix.py --dry-run ``` This creates the virtual environment, installs all dependencies, imports tasks from the registry, applies all defaults, and expands all matrix functions—everything except actually running the evaluations. It's invaluable for verifying that dependencies can be installed, tasks are properly configured, and the exact settings are what you expect. Unlike `flow config` which just parses the config file, `--dry-run` performs the full setup process. To run the config: ```bash flow run matrix.py ``` This will run all 16 evaluations (2 tasks × 2 models × 4 reasoning levels). When complete, you'll find a link to the logs at the bottom of the task results summary. ![Log path printed in terminal](docs/images/logs_terminal.png) To view logs interactively, run: ```bash inspect view --log-dir logs ``` ![Eval logs rendered by Inspect View](docs/images/inspect_view_eval.png) ## Learning More See the following articles to learn more about using Flow: - [Flow Concepts](https://meridianlabs-ai.github.io/inspect_flow/flow_concepts.html): Flow type system, config structure and basics. - [Defaults](https://meridianlabs-ai.github.io/inspect_flow/defaults.html): Define defaults once and reuse them everywhere with automatic inheritance. - [Matrixing](https://meridianlabs-ai.github.io/inspect_flow/matrix.html): Systematic parameter exploration with matrix and with functions. - [Reference](https://meridianlabs-ai.github.io/inspect_flow/reference/): Detailed documentation on the Flow Python API and CLI commands. ## Development To work on development of Inspect Flow, clone the repository and install with the `-e` flag and `[dev, doc]` optional dependencies: ```bash git clone https://github.com/meridianlabs-ai/inspect_flow cd inspect_flow uv sync source .venv/bin/activate ``` Optionally install pre-commit hooks via ```bash make hooks ``` Run linting, formatting, and tests via ```bash make check make test ```
text/markdown
Meridian Labs
null
null
null
MIT License
null
[]
[]
null
null
>=3.10
[]
[]
[]
[ "click>=8.2.1", "deltalake>=0.18.0", "inspect-ai>=0.3.179", "packaging>=21.0", "pyarrow>=17.0.0", "pydantic>=2.11.2", "python-dotenv>=1.1.1", "tomli>=2.0.0; python_version < \"3.11\"", "typing-extensions>=4.9.0", "click; extra == \"doc\"", "griffe; extra == \"doc\"", "jupyter; extra == \"doc\"", "markdown; extra == \"doc\"", "panflute; extra == \"doc\"", "quarto-cli; extra == \"doc\"" ]
[]
[]
[]
[ "Source Code, https://github.com/meridianlabs-ai/inspect_flow", "Issue Tracker, https://github.com/meridianlabs-ai/inspect_flow/issues", "Documentation, https://meridianlabs-ai.github.io/inspect_flow/" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:55:18.073152
inspect_flow-0.4.1.tar.gz
70,408
40/1f/641a294442c02a751b5a53582acfcdb6723b0305551ac4cbb6b1d87db944/inspect_flow-0.4.1.tar.gz
source
sdist
null
false
45b01bfc6a2678ff4a456842865e4c2e
47d196fd2ca2d6dad5f81c7d364ec8acb014d13a5ea49a9757e589c7c591b02e
401f641a294442c02a751b5a53582acfcdb6723b0305551ac4cbb6b1d87db944
null
[ "LICENSE" ]
306
2.4
tklr-dgraham
0.0.55
Reminders Tickler / CLI and Textual UI
<!-- markdownlint-disable MD033 --> <table> <tr> <td style="vertical-align: top; width: 60%;"> <h1>tklr</h1> <p> The term <em>tickler file</em> originally referred to a file system for reminders which used 12 monthly files and 31 daily files. <em>Tklr</em> turns this classic into a local, SQLite-backed reminder system. You enter reminders in plain text; <em>tklr</em> parses dates, recurrence, and metadata as you type, then ranks tasks by urgency and goals by priority. </p> <p>Why try it? <ul> <li>Form‑free entry with live prompts (no forms to fill).</li> <li>CLI and Textual UI with mouse‑free navigation.</li> <li>Multiple reminder types: events, tasks, projects, goals, notes, jots, drafts.</li> <li>Flexible scheduling (fuzzy dates, recurrence, time zones) powered by <em>dateutil</em>.</li> <li>Organized views (Agenda, Next/Last, Queries, bins, hashtags) to surface what matters.</li> </ul></p> <p>If you like fast, local, keyboard‑first tools, <em>tklr</em> gives you a daily brief without a heavyweight app.</p> </td> <td style="width: 40%; vertical-align: middle;"> <figure style="margin: 20px; text-align: center;"> <img src="https://raw.githubusercontent.com/dagraham/tklr-dgraham/master/tklr_logo.avif" alt="tklr logo" title="Tklr" style="max-width: 380px; width: 100%; height: auto;"> <figcaption style="margin-top: 6px; font-style: italic;">Make the most of your time!</figcaption> </figure> </td> </tr> </table> Ready to dive deeper? This introduction is best viewed at [GitHub.io](https://dagraham.github.io/tklr-dgraham/). *Tklr* itself is available from [PyPi](https://pypi.org/project/tklr-dgraham/), the source code from [GitHub](https://github.com/dagraham/tklr-dgraham) and further discussion at [Tklr Discussions](https://github.com/dagraham/tklr-dgraham/discussions). <a id="table-of-contents"></a> <h3>Table of Contents</h3> <details> <summary><strong>Show/Hide</strong></summary> <ul> <li> <details> <summary><a href="#1-what-makes-tklr-different">1. What makes <em>tklr</em> different</a></summary> <ul> <li><a href="#11-form-free-entry">1.1. Form-Free entry</a></li> <li><a href="#12-reminders-to-suit-the-purpose">1.2. Reminders to suit the purpose</a></li> <li><a href="#13-mouse-free-navigation">1.3. Mouse-Free navigation</a></li> <li><a href="#14-agenda-view-your-daily-brief">1.4. Agenda View: Your daily brief</a></li> <li><a href="#15-weeks-next-and-last-views-whats-happening-and-when">1.5. Weeks, Next and Last Views: What's happening and when</a></li> <li><a href="#16-jots-and-jot-uses-views-where-did-the-time-go">1.6. Jots and Jot Uses Views: Where did the time go</a></li> <li><a href="#17-jots-and-gtd">1.7. Jots and GTD</a></li> <li><a href="#18-gtd-and-task-view">1.8. GTD and Task View</a></li> <li><a href="#19-bins-and-hash-tags-views-organizing-your-reminders">1.9. Bins and Hash-Tags Views: Organizing your reminders</a></li> <li><a href="#110-query-and-find-views-wheres-waldo">1.10. Query and Find Views: Where's Waldo</a></li> <li><a href="#111-sqlite3-data-store">1.11. SQLite3 Data Store</a></li> </ul> </details> <details> <summary><a href="#2-details">2. Details</a></summary> <ul> <li><a href="#21-datetimes">2.1. Datetimes</a></li> <li><a href="#22-timedeltas">2.2. TimeDeltas</a></li> <li><a href="#23-scheduled-datetime">2.3. Scheduled datetime</a></li> <li><a href="#24-extent-timedelta">2.4. Extent timedelta</a></li> <li><a href="#25-notice">2.5. Notice</a></li> <li><a href="#26-wrap">2.6. Wrap</a></li> <li><a href="#27-alert">2.7. Alert</a></li> <li><a href="#28-recurrence">2.8. Recurrence</a></li> <li><a href="#29-masked-information">2.9. Masked Information</a></li> <li><a href="#210-hashtags">2.10. HashTags</a></li> <li><a href="#211-anniversaries">2.11. Anniversaries</a></li> <li><a href="#212-timezones">2.12. Timezones</a></li> <li><a href="#213-urgency">2.13. Urgency</a></li> <li><a href="#214-priority">2.14. Priority</a></li> <li><a href="#215-open-with-default">2.15. Open with default</a></li> <li><a href="#216-away-from-your-computer-use-the-cloud">2.16. Away from your computer? Use the cloud</a></li> <li><a href="#217-palette-view-customizing-theme-color-settings">2.17. Palette View: Customizing Theme Color Settings</a></li> </ul> </details> <details> <summary><a href="#3-getting-started">3. Getting Started</a></summary> </details> <details> <summary><a href="#4-using-the-command-line-interface">4. Using the Command Line Interface</a></summary> </details> <details> <summary><a href="#5-coming-from-etm">5. Coming from <em>etm</em></a></summary> </details> <details> <summary><a href="#6-developer-guide">6. Developer Guide</a></summary> </details> </li> </ul> </details> <p>This <em>Table of Contents</em> is expandable with links to all the major sections and with <code>↩︎</code> links throughout the document to return to it.</p> ## 1. What makes tklr different ### 1.1. Form-Free entry <div style="overflow: auto;"> <img src="https://raw.githubusercontent.com/dagraham/tklr-dgraham/master/screenshots/demo.gif" alt="Description" style="float: right; margin-left: 20px; width: 460px; margin-bottom: 10px;"> <p> Rather than filling out fields in a form to create or edit reminders, a simple entry field is provided for text input together with a prompt area which provides <i>instantaneous feedback</i>. </p> <p> This animation shows the entire process for creating a new reminder. The individual entry steps are listed below. <ul> <li>the type character, here an <code>*</code> for an <em>event</em> </li> <li>the subject, "Lunch with Ed" </li> <li>the scheduled time, <code>@s 12p</code> - today is assumed </li> <li>the extent (duration) of the event, <code>@e 1h30m</code> - an hour and thirty minutes </li> <li>an alert, <code>@e 15m: n</code> - fifteen minutes before the event, trigger a notification </li> <li>the resulting event displayed in <em>Agenda View</em> </li> </ul> </p> </div> <div style="clear: both;"></div> [↩︎](#table-of-contents) ### 1.2. Reminders to suit the purpose *tklr* has seven types of reminders, each with a corresponding type character: | item type | character | description | | --------- | :-------: | --------------------------------------- | | event | * | happens at a particular time | | task | ~ | requires an action to complete | | project | ^ | collection of related tasks | | goal | ! | targets action at a specified frequency | | note | % | information for future reference | | jot | - | timestamped message to self | | draft | ? | preliminary/unfinished reminder | Here are some illustrations of how the various types. <a id="reminder-types"></a> <ul> <li><a href="#121-an-event-lunch-with-ed-extended">1.2.1. An <em>event</em>: lunch with Ed (discussed)</a></li> <li><a href="#122-a-task-pick-up-milk">1.2.2. A <em>task</em>: pick up milk</a></li> <li><a href="#123-a-repeating-event-trash-pickup">1.2.3. A <em>repeating event</em>: trash pickup</a></li> <li><a href="#124-an-event-that-repeats-irregularly-dental-appointment">1.2.4. An <em>event that repeats irregularly</em>: dental appointment</a></li> <li><a href="#125-a-complicated-but-regularly-repeating-task-vote-for-president">1.2.5. A <em>complicated</em> but regularly repeating task: vote for president</a></li> <li><a href="#126-an-offset-task-fill-bird-feeders">1.2.6. An <em>offset task</em>: fill bird feeders</a></li> <li><a href="#127-a-note-a-favorite-churchill-quotation">1.2.7. A <em>note</em>: a favorite Churchill quotation</a></li> <li><a href="#128-a-project-build-a-dog-house-with-component-tasks">1.2.8. A <em>project</em>: build a dog house with component tasks</a></li> <li><a href="#129-a-goal-interval-training-3-times-each-week">1.2.9. A <em>goal</em>: interval training 3 times each week</a></li> <li><a href="#1210-a-jot-taking-a-walk">1.2.10. A <em>jot</em>: taking a walk</a></li> <li><a href="#1211-a-draft-reminder-meet-alex-for-coffee---time-to-be-determined">1.2.11. A <em>draft</em> reminder: meet Alex for coffee - time to be determined</a></li> </ul> #### 1.2.1. An _event_: lunch with Ed (extended) <div style="overflow:auto;"> <pre style="float:right; margin-left:20px; width:460px; background:#111; color:#ddd; padding:12px; border-radius:6px;"> <code>* lunch with Ed @s 12p fri @e 1h30m @a 15m: n </code> </pre> <p>The <code>*</code> makes this reminder an <i>event</i> with whatever follows until the next <code>@</code> character as the subject. The <code>@s</code> attribute sets the <i>scheduled</i> or starting time for 12pm on the first Friday on or after today and the <code>@e 1h30m</code> attribute sets the <i>extent</i> for one hour and thirty minutes. This event will thus be displayed as occupying the period <code>12-1:30pm</code> on the day of the event. The distinguishing feature of an <i>event</i> is that it occurs at a particular time and the <code>@s</code> attribute is therefore required. </p> <p>Provided that <em>tklr ui</em> is running, <code>@a 15m: n</code> will trigger a built-in <em>notify</em> alert fifteen minutes before the start of the event which sounds a bell and posts a message on the <em>tklr</em> display showing the subject and time of the event. </p> </div> <div style="clear:both;"></div> #### 1.2.2. A _task_: pick up milk <div style="overflow:auto;"> <pre style="float:right; margin-left:20px; width:460px; background:#111; color:#ddd; padding:12px; border-radius:6px;"> <code>~ pick up milk </code> </pre> <p>The beginning <code>~</code> type character makes this reminder a <i>task</i> with the following <code>pick up milk</code> as the <i>subject</i>. </p> <p>Using an <code>@s</code> attribute is optional and, when specified, it sets the time at which the task should be <strong>completed</strong>, not begun. The <code>@e</code> attribute is also optional and, when given, is intepreted as the estimated time period required for completion. </p> </div> <div style="clear:both;"></div> #### 1.2.3. A _repeating event_: trash pickup <div style="overflow:auto;"> <pre style="float:right; margin-left:20px; width:460px; background:#111; color:#ddd; padding:12px; border-radius:6px;"> <code>* trash pickup @s 8a mon @n 1d @r w &w MO </code> </pre> <p>This <em>event</em> repeats because of the <code>@r w &w MO</code> each week on Mondays. Because of the <code>@n 1d</code> a notice will be posted in <em>Agenda View</em> when the current date is within one day of the scheduled datetime or, in this case, on Sundays. This serves as a reminder to put the trash at the curb before 8am Mondays. Why not use a <em>task</em> for this? A task would require being marked finished each week to avoid accumulating past due instances - even when out of town with neither trash nor opportunity for placement at the curb. </p> </div> <div style="clear:both;"></div> #### 1.2.4. An _event that repeats irregularly_: dental appointment <div style="overflow:auto;"> <pre style="float:right; margin-left:20px; width:460px; background:#111; color:#ddd; padding:12px; border-radius:6px;"> <code>* dental exam and cleaning @s 2p feb 5 @e 45m @+ 9am Sep 3 </code> </pre> <p>This event specifies an appointment for a 45 minute dental exam and cleaning starting at 2pm on February 5 and then again, because of the <code>@+</code> attribute, at 9am on September 3. </p> <p>Need to add another datetime to an existing reminder? Just add an <code>@+</code> attribute with a comma separated list of as many additional dates or datetimes as needed. </p> </div> <div style="clear:both;"></div> #### 1.2.5. A _complicated_ but regularly repeating task: vote for president <div style="overflow:auto;"> <pre style="float:right; margin-left:20px; width:460px; background:#111; color:#ddd; padding:12px; border-radius:6px;"> <code>~ vote for president @s nov 1 2020 @r y &i 4 &w TU &d 2, 3, 4, 5, 6, 7, 8 &m 11 </code> </pre> <p>Here is another, more complicated, but still <em>regularly repeating</em> reminder. Beginning with November, 2020, this <em>task</em> repeats every 4 years on the first Tuesday after a Monday in November (a <em>Tuesday</em> whose <em>month day</em> falls between 2 and 8 in the 11th <em>month</em>). </p> <p>This is a good illustration of the power of the <em>dateutil</em> library. Note that the only role of <code>@s nov 1 2020</code> is to limit the repetitions generated by <code>@r</code> to those falling on or after November 1, 2020 and occur on that year or a multiple of 4 years after that year. </p> </div> <div style="clear:both;"></div> #### 1.2.6. An _offset task_: fill bird feeders <div style="overflow:auto;"> <pre style="float:right; margin-left:20px; width:460px; background:#111; color:#ddd; padding:12px; border-radius:6px;"> <code>~ fill birdfeeders @s 3p sat @n 1d @o 12d </code> </pre> <p>Because of the <code>@o 12d</code> <em>offset</em> attribute, when this task is completed the <code>@s</code> <em>scheduled</em> datetime will automatically reset to the datetime that falls precisely 12 days after the completion datetime. Whether they are filled early or late, they will still need to be refilled 12 days after they were last filled. Because of the <code>@n 1d</code> <em>notice</em> attribute, this task will <em>not</em> appear in the <em>Agenda View</em> task list until the the current datetime is within one day of the <em>scheduled</em> datetime. </p> </div> <div style="clear:both;"></div> Since the <code>@o</code> attribute involves resetting attributes in a way that effectively repeats the <em>task</em>: 1. `@o` can only be used with _tasks_ 2. Using `@o` precludes the use of `@r` It is worth noting the different roles of two attributes in events and tasks. 1. The <em>scheduled</em> datetime attribute describes when an event begins but when a task should be completed. 2. The <em>notice</em> attribute provides an early warning for an event but postpones the disclosure of a task. #### 1.2.7. A _note_: a favorite Churchill quotation <div style="overflow:auto;"> <pre style="float:right; margin-left:20px; width:460px; background:#111; color:#ddd; padding:12px; border-radius:6px;"> <code>% Give me a pig - #Churchill @d Dogs look up at you. Cats look down at you. Give me a pig - they look you in the eye and treat you as an equal. @b quotations </code> </pre> <p>The beginning <code>%</code> makes this reminder a <i>note</i> with the <i>subject</i>, <code>Give me a pig - #Churchill</code>. The optional <i>details</i> attribute follows the <code>@d</code> and is meant to be more expansive - analogous to the body of an email. The hash character that precedes 'Churchill' in the subject makes that word a <i>hash tag</i> for listing in <i>Tags View</i>. The <code>@b</code> entry adds this reminder to the 'quotations' <i>bin</i> for listing in <i>Bins View</i>. </p> </div> <div style="clear:both;"></div> <a id="doghouse-example"></a> #### 1.2.8. A _project_: build a dog house with component tasks <div style="overflow:auto;"> <pre style="float:right; margin-left:20px; width:460px; background:#111; color:#ddd; padding:12px; border-radius:6px;"> <code>^ Build dog house @~ pick up materials &r 1 &e 4h @~ cut pieces &r 2: 1 &e 3h @~ assemble &r 3: 2 &e 2h @~ sand &r 4: 3 &e 1h @~ paint &r 5: 4 &e 4h </code> </pre> <p>The beginning <code>^</code> makes this a <i>project</i>. This is a collection of related tasks specified by the <code>@~</code> entries. In each task, the <code>&r X: Y</code> <em>requires</em> attribute sets <code>X</code> as the label for the task and sets the task labeled <code>Y</code> as a requirement or prerequisite for <code>X</code>. E.g., <code>&r 3: 2</code> establishes "3" as the label for assemble and "2" (cut pieces) as a prerequisite. The <code>&e</code> <i>extent</i> entries give estimates of the times required to complete the various tasks. </p> </div> <div style="clear:both;"></div> #### 1.2.9. A _goal_: interval training 3 times each week <div style="overflow:auto;"> <pre style="float:right; margin-left:20px; width:460px; background:#111; color:#ddd; padding:12px; border-radius:6px;"> <code>! interval training @s 2026-01-05 @o 3/1w </code> </pre> <p>The beginning <code>!</code> type character makes this reminder a <i>goal</i> with the following <code>interval training</code> as the <i>subject</i>. The <code>@t 3/1w</code> attribute is required and sets the <i>target</i> to be 3 completions during the period of one week starting at midnight on '2026-01-05', because of the <code>@s</code> attribute, and ending one week later at midnight on '2026-01-12', because of the '1w' target period. </p> </div> <div style="clear:both;"></div> When a *goal* is created, the attribute `@k 0` is automatically added to indicate that the current *completion count* is zero. When a completion is recorded for the *goal*, this count is automatically increased by one. This process continues until 1. the period allowed for completing the goal expires or 2. the completion count reaches the target number of completions In either case, `@k` is reset to zero and `@s` is reset to the previous value *plus* the period allowed for completion of the goal, i.e, to the *end* of the period originally allowed for completion. #### 1.2.10. A _jot_: taking a walk Tracking where a resource goes is the key to managing any scarce resource - your time is no exception. A *jot* is a reminder type designed to facilitate this purpose. It provides a way of *quickly* recording a *jot* of time-stamped information as a *message to self*. It is sufficiently different from the other reminder types to warrant some discussion before giving an example. Imagine that *tklr* is running on your computer and that, in the midst of your hectic day, you could reach over, press "+" to create a new reminder, enter "-" to make it a *jot* and follow with the *subject* - a *brief* phrase - just enough to trigger your memory later. Then press Ctrl+S to save - an automatic timestamp will be added. What might you do with these *jots*? As is, these jots can provide a record of what you were doing and when. At the cost of a few seconds per *jot* you can have a daily record of when and what you were doing or thinking. Press "J" to see a list of all your *jots* grouped by week and weekday and, as with all other reminder views in *tklr*, tagged with lower case letters, a, b, c, .... for easy access. When you have time, you might want to: - flush out the *subject* or add a `@d` *details* entry to provide extra detail. - record the time spent. You could, e.g., add `@e 1h15`, to indicate that an hour and fifteen minutes of your precious time was spent on whatever you were doing when the *jot* was recorded. - record the particular <em>use</em> to which the *jot* applies, e.g., `@u exercise.walking` to indicate that this time should be attributed to the *use* "exercise.walking". Press "U" whenever you like to see a report of your *Jot Uses* with *totals* of your time spent by *month* and *use*. - add a hash-tag to the subject or the details of a *jot* to make it easy to find in the *Hash-Tags View*. - convert it to another type of reminder. E.g. `- book a lunch reservation for Friday` might be converted to the event `* lunch with Ed @s fri 1pm` when you make the reservation. Underneath the details is a very simple idea - in the heat of battle when every second counts, *jot* down just enough to trigger your memory later when things have calmed down. <div style="overflow:auto;"> <pre style="float:right; margin-left:20px; width:460px; background:#111; color:#ddd; padding:12px; border-radius:6px;"> <code>- taking a walk @s 2026-01-05 13:14 @e 1h15m @u exercise.walking </code> </pre> <p>Here's an example. The beginning <code>-</code> type character makes this reminder a <i>jot</i> with the following <code>taking a walk</code> as the <i>subject</i>. This much was <em>jotted down</em> before the walk and the <code>@s 2026-01-05 13:14</code> was automatically appended at that time. </p> <p>Later in the day, after the walk was finished and the <em>jots</em> for the day were being reviewed, the <em>time spent</em>, <code>@e 1h15m</code>, and the <em>use</em>, <code>@u exercise.walking</code>, were added. </p> <p>Note that <code>exercise.walking</code> would serve to differentiate this form of exercise from, say, <code>exercise.interval_training</code>, and thus avoid adding "apples and oranges" when reporting the <em>use</em> totals. </p> </div> <div style="clear:both;"></div> #### 1.2.11. A _draft_ reminder: meet Alex for coffee - time to be determined <div style="overflow:auto;"> <pre style="float:right; margin-left:20px; width:460px; background:#111; color:#ddd; padding:12px; border-radius:6px;"> <code>? Coffee with Alex @s fri @e 1h </code> </pre> <p>The beginning <code>?</code> type character makes this a <i>draft</i> reminder. This can be changed to an event when the details are confirmed by replacing the <code>?</code> with an <code>*</code> and adding the time to <code>@s fri</code>. </p> <p> This is a reminder that is not yet finished and, in almost every respect, will be ignored by <em>tklr</em>. The exception is that it will appear highlighted on the current day in <em>Agenda View</em> until it is revised. It can be changed to an <em>event</em> when the details are confirmed by replacing the <code>?</code> with an <code>*</code> and adding the time to <code>@s</code>. </p> </div> <div style="clear:both;"></div> [↩︎](#table-of-contents) ### 1.3. Mouse-Free navigation Each of the main views in *tklr* can be opened by pressing a single key - the first letter of the view's name. | View | Key | Displays | | ---------------- | :---: | ----------------------------------------------- | | Agenda | A | events, goals, tasks | | Bins | B | Tree view of Bins | | Completions | C | Completion datetimes for completed tasks | | Find | F | Case insensitive search in subjects and details | | Goals | G | Case insensitive search in subjects and details | | Hash-Tags | H | List reminders with tags grouped by tag | | Jots | J | Jots by week and weekday | | Last | L | The last instance of reminders before today | | Modified | M | All reminders sorted by the modified timestamp | | Next | N | The next instance of reminders after today | | Query | Q | List matches for a specified query | | Remaining Alerts | R | List remaining alerts for the today | | Jot Uses | U | Jots with totals by month and use | | Weeks | W | Scheduled reminders by week and weekday | Each of these views displays a vertical list of reminders, with each reminder row beginning with a tag from "a", "b", ..., "z", followed by the pertinent details of the reminder including its subject. When necessary, lists are split into pages so that no more than 26 reminders appear on any one page and the left and right cursor keys are used to move back and forth between pages. *The view keys and the list tags are the key to navigating tklr.* On any page, pressing the key corresponding to a tag will open a display with all the details of the corresponding reminder. This is worth emphasizing. *You don't need the cursor keys or the mouse to select a reminder - just press the key corresponding to its tag.* When the details of reminder are being displayed, pressing `enter` will open a menu of various commands applicable to the selected reminder, pressing the key corresponding to the tag of another reminder will switch the details display to that reminder or pressing the upper case letter corresponding to another view will switch to that view. Everything you might want to do to a reminder, to edit, finish, reschedule, delete or whatever is available using these steps: 1. press the key corresponding to the tag of the reminder you want to select 2. press `enter` to open the menu of commands for the selected reminder 3. press the first letter (any case) of the desired command or `escape` to cancel and close the commands menu [↩︎](#table-of-contents) ### 1.4. Agenda View: Your daily brief *Agenda view* displays 1. The next few days of <em>events</em> beginning with today 2. Active <em>goals</em> ordered by their *priority* 3. Available <em>tasks</em> ordered by their *urgency* Times are displayed in the screenshots using _24-hour_ notation. An option can be set to display times using _am/pm_ notation instead. <div style="overflow: auto;"> <img src="https://raw.githubusercontent.com/dagraham/tklr-dgraham/master/screenshots/agenda_screenshot.svg" alt="Description" style="float: right; margin-left: 20px; width: 460px; margin-bottom: 10px;"> <p>While the listing of events begins with the current day, any all-day events or events whose ending times have already passed such as the one tagged <em>a</em> will be dimmed. Additionally an event, such as the one tagged <em>b</em> whose active period overlaps the current moment will be highlighted. </p> <p>The first day of events will always include any <em>notices</em> of upcoming events or <em>draft</em> reminders needing completion in addition to any scheduled events for today. In this case the reminder tagged <em>d</em> indicates that there is an event beginning in 5 days (<code>+5d</code>) with a subject beginning with "Amet porro ..." and a <em>notice attribute</em>, <code>@n x</code>, in which <code>x > 5d</code>. This attribute is the reason this notice of the event is being displayed before its scheduled datetime - it will continue to be displayed on the first day (current date) of Agenda View each day until the day of the event. </p> <p>There is also a draft entry tagged <em>e</em> and displayed in red. This is simply a reminder whose item type is <code>?</code>. This is used to flag a reminder as incomplete as would be the case, e.g., if a final datetime for the event had not yet been finalized. Draft reminders are also displayed on the current, first day in Agenda view until the item type is changed. </p> <p>The list for *goals* includes all goals which have not been completed on the current date, sorted and color coded by their *priority*, which is listed in the first column after the tags. The details for precisely how *priority* is calculated will be described later but the basic idea involves comparing</p> <ol> <li>the rate at which completions would currently need to occur to complete the goal</li> <li>the rate originally specified in the goal</li> </ol> <p>The higher the current rate relative to the original, the higher the *priority*.</p> <p>The list for *tasks* includes all tasks with the possible exception of tasks with both an `@s` (specifying a *due datetime*) and an `@n` entry (specifying a *notification period*). Suppose, for example, that `@s 2026-01-30` and `@n 2d`. The role of these combined entries is to say that the task needs to be finished by `2026-01-30` but you don't want to be bothered about it until two days before that date. This task won't appear in the list until `2026-01-28`.</p> <p>Tasks are sorted by their *urgency*. This calculation is fairly complicated and will be described later. Many factors are involved including the priority of the task, its due datetime, how many tags it has, whether it has a details attribute and so forth. The *weights* attached to these and other characteristics are options which can be set in the user configuration file.</p> <p>*Agenda* is the default view and represents the place to go for what you need to know right now.</p> </div> <div style="clear: both;"></div> [↩︎](#table-of-contents) ### 1.5. Weeks, Next and Last Views: What's happening and when *Weeks View* is dedicated to displaying each instance of your scheduled reminders one week at a time with a *busy bar* at the top to show the busy days during the week at a glance followed by a day by day listing of the scheduled reminders. Two supporting views are limited to displaying a single instance of each scheduled reminder. *Next View*, bound to <code>N</code>, lists the *first* instance occurring on or after the current date in *ascending* order and *Last View*, bound to <code>L</code>, lists the most recent instance occurring *before* the current date in *descending* order. When did you last have your car serviced? *Last View* is the place to look. When is your next dental appointment? *Next View* has the answer. <div style="overflow: auto;"> <img src="https://raw.githubusercontent.com/dagraham/tklr-dgraham/master/screenshots/weeks_screenshot.svg" alt="Description" style="float: right; margin-left: 20px; width: 460px; margin-bottom: 10px;"> <p>Press <code>W</code> to open <em>Weeks View</em> on the current week or press <code>D</code> and enter a date to open the view on the week containing that date. The header displays the date range, year and week number for the displayed week. Left and right cursor keys shift the displayed week backward or forward by one week. Pressing the shift key at the same time increases the shift from one to four weeks. Pressing the space key will jump back to the current week. </p> <p>As with the other tagged views, pressing the key corresponding to the tag of a reminder opens a panel with the details of that reminder. In this case, the details for tag <em>j</em> are being displayed. </p> <p> The <em>busy bar</em> underneath the header provides graphical illustration of the busy times for <em>events</em> during the week. The area under each weekday name has spaces for five blocks. The first (furthest to the left) will be colored orange if one or more <em>all day</em> events are scheduled for that day. The next four blocks correspond to the four 6-hour periods during the day beginning with 00:00 - 05:59 and ending with 18:00 - 23:59 - night, morning, afternoon and evening. </p> <p>The block corresponding to a period will be green if the scheduled time for an event occupies any part of the period. E.g., a single event scheduled for 05:00 - 07:00 would cause both the first and second blocks for that day to be colored green. A block is changed from green to red if the busy periods for two or more events overlap and thus <em>conflict</em>. The red block for Tuesday, e.g., reflects the conflict during the period 11:00 - 11:15 by the events tagged <em>b</em> and <em>c</em>. </p> <p>Note that only <em>events</em> with an <em>extent</em> contribute to the <em>busy bar</em>. E.g., the <em>event</em> tagged <em>i</em> on Friday has no extent and thus no effect on the <em>busy bar</em> "morning" slot for that day. Similarly, the <em>task</em> tagged <em>j</em> whose details are displayed, is scheduled for 10:15 - 11:15 and yet, being a task, also has no effect on that "morning" slot. </p> </div> <div style="clear: both;"></div> [↩︎](#table-of-contents) ### 1.6. Jots and Jot Uses Views: Where did the time go These screenshots reflect a configuration setting that rounds reported <code>@e</code> times up to the next integer multiple of <code>6</code> minutes and thus reports times in hours and tenths. <div style="overflow: auto;"> <img src="https://raw.githubusercontent.com/dagraham/tklr-dgraham/master/screenshots/jots_jots.svg" alt="Description" style="float: right; margin-left: 20px; width: 460px; margin-bottom: 10px;"> <p> <em>Jots View</em> is similar to <em>Weeks View</em> - reminders are grouped by <em>week</em> and then <em>week day</em> with the same key bindings used for navigation. The <em>jots</em> are displayed in different colors for <ul> <li>those with neither <code>@e</code> nor <code>@u</code> entries</li> <li>those with only <code>@e</code> entries</li> <li>those with only <code>@u</code> entries</li> <li>those with both <code>@e</code> and <code>@u</code> entries</li> </ul> When either or both of these attributes are present, they are given in parentheses after the <em>subject</em>. </p> </div> <div style="clear: both;"></div> <div style="overflow: auto;"> <img src="https://raw.githubusercontent.com/dagraham/tklr-dgraham/master/screenshots/jots_uses.svg" alt="Description" style="float: right; margin-left: 20px; width: 460px; margin-bottom: 10px;"> <p> <em>Jots Uses View</em> groups by <em>month</em> and <em>use</em>. <em>Jots</em> without an entry for <code>@u</code> are grouped under <em>unassigned</em>. </p> <p> The listing for each <em>jot</em> gives <ol> <li>tag</li> <li><code>@s entry</code> hours:minutes</li> <li><code>@s entry</code> month day</li> <li><code>@e entry</code></li> in hours and tenths if given else blank <li><em>subject</em></li> </ol> All <em>jots</em> are listed using the same colors as were used in <em>Jots View</em>. </p> </div> <div style="clear: both;"></div> <div style="overflow:auto;"> <pre style="float:right; margin-left:20px; width:460px; background:#111; color:#ddd; padding:12px; border-radius:6px;"> <code>% tklr uses report --use exercise Jot Uses - Jan 2026 - Feb 2026: 3.9h Jan 2026: 0.5h exercise.bike: 0.5h 14:13 26 0.5h Ut tempora consectetur exercise.walking 12:30 31 Ut porro dolor non ut Feb 2026: 3.4h exercise.bike: 1.6h 14:13 2 Voluptatem aliquam ipsum velit 11:00 6 Ipsum ipsum est 11:15 6 1.3h Dolor dolorem labore sed 13:15 8 0.3h Quaerat etincidunt quisquam dolor exercise.walking: 1.8h 13:30 2 Magnam quaerat dolor non 12:45 5 1.8h Porro est sit 13:15 5 Adipisci tempora neque 15:15 5 Adipisci sed voluptatem sit porro 15:30 5 Dolorem sit dolore non dolorem </code> </pre> <p>The CLI version is similar. This one using the command <code> tklr uses report --use exercise</code> limits the <em>uses</em> to those containing a match for "exercise". </p> </div> <div style="clear:both;"></div> <div style="overflow:auto;"> <pre style="float:right; margin-left:20px; width:460px; background:#111; color:#ddd; padding:12px; border-radius:6px;"> <code>% tklr uses report --use meditation --verbose Jot Uses - Jan 2026 - Feb 2026: 1.0h Jan 2026 meditation 14:13 26 Dolore aliquam consectetur u without e #lorem Feb 2026: 1.0h meditation: 1.0h 14:13 2 0.5h Quaerat numquam eius amet u and e #lorem 12:30 5 Porro est sed Porro amet quisquam eius amet labore dolor. Ut labore ut quaerat dolorem magnam quiquia. Quisquam non est quisquam dolor neque tempora velit. Dolore ut numquam sit velit aliquam ipsum. #lorem #amber 14:34 9 0.5h Dolorem quaerat quaerat consectetur u and e #lorem </code> </pre> <p>This one shows <em>uses</em> matching "meditation" and, because of the <code>--verbose</code> argument, also displays the <code>@d</code> <em>details</em> attribute indented under the subject. </p> </div> <div style="clear:both;"></div> [↩︎](#table-of-contents) ### 1.7. Jots and GTD <div style="overflow: auto;"> <img src="https://raw.githubusercontent.com/dagraham/tklr-dgraham/master/screenshots/GTD.png" alt="Description" style="float: right; margin-left: 20px; width: 460px; margin-bottom: 10px;"> <p> <em>Jots View</em> is similar to <em>Weeks View</em> - reminders are grouped by <em>week</em> and then <em>week day</em> with the same key bindings used for navigation. The <em>jots</em> are displayed in different colors for <ul> <li>those with neither <code>@e</code> nor <code>@u</code> entries</li> <li>those with only <code>@e</code> entries</li> <li>those with only <code>@u</code> entries</li> <li>those with both <code>@e</code> and <code>@u</code> entries</li> </ul> When either or both of these attributes are present, they are given in parentheses after the <em>subject</em>. </p> </div> <div style="clear: both;"></div> [↩︎](#table-of-contents) ### 1.8. GTD and Task View [↩︎](#table-of-contents) ### 1.9. Bins and Hash-Tags Views: Organizing your reminders *Tklr* provides two complementary methods for organizing your reminders: <ol> <li>Using the attribute <code>@b</code> to attach the the name of a bin to a reminder and the related <em>Bins View</em></li> <li>Using a <em>hash-tag</em>, i.e., <code>&#x23;</code> followed without spaces by an arbitrary word, in either the <em>subject</em> or the <em>details</em> attribute of reminders and the related <em>Tags View</em></li> </ol> #### 1.9.1 Bins View The _Bins View_ displays a hierarchical, tree view of _bins_ and _reminders_. Think of _bins_ as directories, _reminders_ as files and _Bin View_ as a file browser. The main difference is that _reminders_ can belong to more than one _bin_ or to none at all. As an illustration of the power of being able to place a reminder in many bins consider a note describing a visit to Lille, France on November 11, 2025 which involved meeting a dear friend, Mary Smith for lunch. This note might belong to all of these bins: - _travel_ (in _activities_) - _2025:11_ (in _journal_) - _Mary Smith_ (in _people:S_) - _Lille_ (in _places:France_) Many note taking applications provide a means for establishing links between notes. The terms _Zettelkasten_ and _Second Brain_ come to mind. A different approach is taken in _tklr_ where _bins_ serve as containers for both reminders and other bins. While a system of links between reminders might be broken by the removal of a reminder, when a reminder is removed from _tklr_, it simply disappears from the relevant bin membership lists. Bins themselves and their membership lists are otherwise unaffected. <div style="overflow: auto;"> <img src="https://raw.githubusercontent.com/dagraham/tklr-dgraham/master/screenshots/bin_root_screenshot.svg" alt="Description" style="float: right; margin-left: 20px; width: 460px; margin-bottom: 10px;"> <p>These are the important facts about <em>Bins</em>: </p> <ul> <li>Bin names are unique</li> <li>A bin can contain many other bins (children)</li> <li>A bin can belong to at most one other bin (parent)</li> <li>A reminder can belong to one or more bins by adding an <code>@b NAME</code> attribute with a unique <em>NAME</em> for each</li> </ul> <p>This is the opening, root level in <em>Bins view</em>. </p> </div> <div style="clear: both;"></div> <div style="overflow: auto;"> <img src="https://raw.githubusercontent.com/dagraham/tklr-dgraham/master/screenshots/bin_library_screenshot.svg" alt="Description" style="float: right; margin-left: 20px; width: 460px; margin-bottom: 10px;"> <p>Press <em>c</em> to open the <em>library</em> bin with its tagged list of children which now includes both bins and reminders. </p> <ul> <li>Use the tag for a bin to open the bin </li>
text/markdown
null
Daniel Graham <dnlgrhm@gmail.com>
null
null
This project is licensed under the GNU General Public License v3.0 or later. ----------------------------------------------------------------------- GNU GENERAL PUBLIC LICENSE Version 3, 29 June 2007 Copyright (C) 2007 Free Software Foundation, Inc. <http://fsf.org/> Everyone is permitted to copy and distribute verbatim copies of this license document, but changing it is not allowed. Preamble The GNU General Public License is a free, copyleft license for software and other kinds of works. The licenses for most software and other practical works are designed to take away your freedom to share and change the works. By contrast, the GNU General Public License is intended to guarantee your freedom to share and change all versions of a program--to make sure it remains free software for all its users. We, the Free Software Foundation, use the GNU General Public License for most of our software; it applies also to any other work released this way by its authors. You can apply it to your programs, too. When we speak of free software, we are referring to freedom, not price. Our General Public Licenses are designed to make sure that you have the freedom to distribute copies of free software (and charge for them if you wish), that you receive source code or can get it if you want it, that you can change the software or use pieces of it in new free programs, and that you know you can do these things. To protect your rights, we need to prevent others from denying you these rights or asking you to surrender the rights. Therefore, you have certain responsibilities if you distribute copies of the software, or if you modify it: responsibilities to respect the freedom of others. For example, if you distribute copies of such a program, whether gratis or for a fee, you must pass on to the recipients the same freedoms that you received. You must make sure that they, too, receive or can get the source code. And you must show them these terms so they know their rights. Developers that use the GNU GPL protect your rights with two steps: (1) assert copyright on the software, and (2) offer you this License giving you legal permission to copy, distribute and/or modify it. For the developers' and authors' protection, the GPL clearly explains that there is no warranty for this free software. For both users' and authors' sake, the GPL requires that modified versions be marked as changed, so that their problems will not be attributed erroneously to authors of previous versions. Some devices are designed to deny users access to install or run modified versions of the software inside them, although the manufacturer can do so. This is fundamentally incompatible with the aim of protecting users' freedom to change the software. The systematic pattern of such abuse occurs in the area of products for individuals to use, which is precisely where it is most unacceptable. Therefore, we have designed this version of the GPL to prohibit the practice for those products. If such problems arise substantially in other domains, we stand ready to extend this provision to those domains in future versions of the GPL, as needed to protect the freedom of users. Finally, every program is threatened constantly by software patents. States should not allow patents to restrict development and use of software on general-purpose computers, but in those that do, we wish to avoid the special danger that patents applied to a free program could make it effectively proprietary. To prevent this, the GPL assures that patents cannot be used to render the program non-free. The precise terms and conditions for copying, distribution and modification follow. TERMS AND CONDITIONS 0. Definitions. "This License" refers to version 3 of the GNU General Public License. "Copyright" also means copyright-like laws that apply to other kinds of works, such as semiconductor masks. "The Program" refers to any copyrightable work licensed under this License. Each licensee is addressed as "you". "Licensees" and "recipients" may be individuals or organizations. To "modify" a work means to copy from or adapt all or part of the work in a fashion requiring copyright permission, other than the making of an exact copy. The resulting work is called a "modified version" of the earlier work or a work "based on" the earlier work. A "covered work" means either the unmodified Program or a work based on the Program. To "propagate" a work means to do anything with it that, without permission, would make you directly or secondarily liable for infringement under applicable copyright law, except executing it on a computer or modifying a private copy. Propagation includes copying, distribution (with or without modification), making available to the public, and in some countries other activities as well. To "convey" a work means any kind of propagation that enables other parties to make or receive copies. Mere interaction with a user through a computer network, with no transfer of a copy, is not conveying. An interactive user interface displays "Appropriate Legal Notices" to the extent that it includes a convenient and prominently visible feature that (1) displays an appropriate copyright notice, and (2) tells the user that there is no warranty for the work (except to the extent that warranties are provided), that licensees may convey the work under this License, and how to view a copy of this License. If the interface presents a list of user commands or options, such as a menu, a prominent item in the list meets this criterion. 1. Source Code. The "source code" for a work means the preferred form of the work for making modifications to it. "Object code" means any non-source form of a work. A "Standard Interface" means an interface that either is an official standard defined by a recognized standards body, or, in the case of interfaces specified for a particular programming language, one that is widely used among developers working in that language. The "System Libraries" of an executable work include anything, other than the work as a whole, that (a) is included in the normal form of packaging a Major Component, but which is not part of that Major Component, and (b) serves only to enable use of the work with that Major Component, or to implement a Standard Interface for which an implementation is available to the public in source code form. A "Major Component", in this context, means a major essential component (kernel, window system, and so on) of the specific operating system (if any) on which the executable work runs, or a compiler used to produce the work, or an object code interpreter used to run it. The "Corresponding Source" for a work in object code form means all the source code needed to generate, install, and (for an executable work) run the object code and to modify the work, including scripts to control those activities. However, it does not include the work's System Libraries, or general-purpose tools or generally available free programs which are used unmodified in performing those activities but which are not part of the work. For example, Corresponding Source includes interface definition files associated with source files for the work, and the source code for shared libraries and dynamically linked subprograms that the work is specifically designed to require, such as by intimate data communication or control flow between those subprograms and other parts of the work. The Corresponding Source need not include anything that users can regenerate automatically from other parts of the Corresponding Source. The Corresponding Source for a work in source code form is that same work. 2. Basic Permissions. All rights granted under this License are granted for the term of copyright on the Program, and are irrevocable provided the stated conditions are met. This License explicitly affirms your unlimited permission to run the unmodified Program. The output from running a covered work is covered by this License only if the output, given its content, constitutes a covered work. This License acknowledges your rights of fair use or other equivalent, as provided by copyright law. You may make, run and propagate covered works that you do not convey, without conditions so long as your license otherwise remains in force. You may convey covered works to others for the sole purpose of having them make modifications exclusively for you, or provide you with facilities for running those works, provided that you comply with the terms of this License in conveying all material for which you do not control copyright. Those thus making or running the covered works for you must do so exclusively on your behalf, under your direction and control, on terms that prohibit them from making any copies of your copyrighted material outside their relationship with you. Conveying under any other circumstances is permitted solely under the conditions stated below. Sublicensing is not allowed; section 10 makes it unnecessary. 3. Protecting Users' Legal Rights From Anti-Circumvention Law. No covered work shall be deemed part of an effective technological measure under any applicable law fulfilling obligations under article 11 of the WIPO copyright treaty adopted on 20 December 1996, or similar laws prohibiting or restricting circumvention of such measures. When you convey a covered work, you waive any legal power to forbid circumvention of technological measures to the extent such circumvention is effected by exercising rights under this License with respect to the covered work, and you disclaim any intention to limit operation or modification of the work as a means of enforcing, against the work's users, your or third parties' legal rights to forbid circumvention of technological measures. 4. Conveying Verbatim Copies. You may convey verbatim copies of the Program's source code as you receive it, in any medium, provided that you conspicuously and appropriately publish on each copy an appropriate copyright notice; keep intact all notices stating that this License and any non-permissive terms added in accord with section 7 apply to the code; keep intact all notices of the absence of any warranty; and give all recipients a copy of this License along with the Program. You may charge any price or no price for each copy that you convey, and you may offer support or warranty protection for a fee. 5. Conveying Modified Source Versions. You may convey a work based on the Program, or the modifications to produce it from the Program, in the form of source code under the terms of section 4, provided that you also meet all of these conditions: a) The work must carry prominent notices stating that you modified it, and giving a relevant date. b) The work must carry prominent notices stating that it is released under this License and any conditions added under section 7. This requirement modifies the requirement in section 4 to "keep intact all notices". c) You must license the entire work, as a whole, under this License to anyone who comes into possession of a copy. This License will therefore apply, along with any applicable section 7 additional terms, to the whole of the work, and all its parts, regardless of how they are packaged. This License gives no permission to license the work in any other way, but it does not invalidate such permission if you have separately received it. d) If the work has interactive user interfaces, each must display Appropriate Legal Notices; however, if the Program has interactive interfaces that do not display Appropriate Legal Notices, your work need not make them do so. A compilation of a covered work with other separate and independent works, which are not by their nature extensions of the covered work, and which are not combined with it such as to form a larger program, in or on a volume of a storage or distribution medium, is called an "aggregate" if the compilation and its resulting copyright are not used to limit the access or legal rights of the compilation's users beyond what the individual works permit. Inclusion of a covered work in an aggregate does not cause this License to apply to the other parts of the aggregate. 6. Conveying Non-Source Forms. You may convey a covered work in object code form under the terms of sections 4 and 5, provided that you also convey the machine-readable Corresponding Source under the terms of this License, in one of these ways: a) Convey the object code in, or embodied in, a physical product (including a physical distribution medium), accompanied by the Corresponding Source fixed on a durable physical medium customarily used for software interchange. b) Convey the object code in, or embodied in, a physical product (including a physical distribution medium), accompanied by a written offer, valid for at least three years and valid for as long as you offer spare parts or customer support for that product model, to give anyone who possesses the object code either (1) a copy of the Corresponding Source for all the software in the product that is covered by this License, on a durable physical medium customarily used for software interchange, for a price no more than your reasonable cost of physically performing this conveying of source, or (2) access to copy the Corresponding Source from a network server at no charge. c) Convey individual copies of the object code with a copy of the written offer to provide the Corresponding Source. This alternative is allowed only occasionally and noncommercially, and only if you received the object code with such an offer, in accord with subsection 6b. d) Convey the object code by offering access from a designated place (gratis or for a charge), and offer equivalent access to the Corresponding Source in the same way through the same place at no further charge. You need not require recipients to copy the Corresponding Source along with the object code. If the place to copy the object code is a network server, the Corresponding Source may be on a different server (operated by you or a third party) that supports equivalent copying facilities, provided you maintain clear directions next to the object code saying where to find the Corresponding Source. Regardless of what server hosts the Corresponding Source, you remain obligated to ensure that it is available for as long as needed to satisfy these requirements. e) Convey the object code using peer-to-peer transmission, provided you inform other peers where the object code and Corresponding Source of the work are being offered to the general public at no charge under subsection 6d. A separable portion of the object code, whose source code is excluded from the Corresponding Source as a System Library, need not be included in conveying the object code work. A "User Product" is either (1) a "consumer product", which means any tangible personal property which is normally used for personal, family, or household purposes, or (2) anything designed or sold for incorporation into a dwelling. In determining whether a product is a consumer product, doubtful cases shall be resolved in favor of coverage. For a particular product received by a particular user, "normally used" refers to a typical or common use of that class of product, regardless of the status of the particular user or of the way in which the particular user actually uses, or expects or is expected to use, the product. A product is a consumer product regardless of whether the product has substantial commercial, industrial or non-consumer uses, unless such uses represent the only significant mode of use of the product. "Installation Information" for a User Product means any methods, procedures, authorization keys, or other information required to install and execute modified versions of a covered work in that User Product from a modified version of its Corresponding Source. The information must suffice to ensure that the continued functioning of the modified object code is in no case prevented or interfered with solely because modification has been made. If you convey an object code work under this section in, or with, or specifically for use in, a User Product, and the conveying occurs as part of a transaction in which the right of possession and use of the User Product is transferred to the recipient in perpetuity or for a fixed term (regardless of how the transaction is characterized), the Corresponding Source conveyed under this section must be accompanied by the Installation Information. But this requirement does not apply if neither you nor any third party retains the ability to install modified object code on the User Product (for example, the work has been installed in ROM). The requirement to provide Installation Information does not include a requirement to continue to provide support service, warranty, or updates for a work that has been modified or installed by the recipient, or for the User Product in which it has been modified or installed. Access to a network may be denied when the modification itself materially and adversely affects the operation of the network or violates the rules and protocols for communication across the network. Corresponding Source conveyed, and Installation Information provided, in accord with this section must be in a format that is publicly documented (and with an implementation available to the public in source code form), and must require no special password or key for unpacking, reading or copying. 7. Additional Terms. "Additional permissions" are terms that supplement the terms of this License by making exceptions from one or more of its conditions. Additional permissions that are applicable to the entire Program shall be treated as though they were included in this License, to the extent that they are valid under applicable law. If additional permissions apply only to part of the Program, that part may be used separately under those permissions, but the entire Program remains governed by this License without regard to the additional permissions. When you convey a copy of a covered work, you may at your option remove any additional permissions from that copy, or from any part of it. (Additional permissions may be written to require their own removal in certain cases when you modify the work.) You may place additional permissions on material, added by you to a covered work, for which you have or can give appropriate copyright permission. Notwithstanding any other provision of this License, for material you add to a covered work, you may (if authorized by the copyright holders of that material) supplement the terms of this License with terms: a) Disclaiming warranty or limiting liability differently from the terms of sections 15 and 16 of this License; or b) Requiring preservation of specified reasonable legal notices or author attributions in that material or in the Appropriate Legal Notices displayed by works containing it; or c) Prohibiting misrepresentation of the origin of that material, or requiring that modified versions of such material be marked in reasonable ways as different from the original version; or d) Limiting the use for publicity purposes of names of licensors or authors of the material; or e) Declining to grant rights under trademark law for use of some trade names, trademarks, or service marks; or f) Requiring indemnification of licensors and authors of that material by anyone who conveys the material (or modified versions of it) with contractual assumptions of liability to the recipient, for any liability that these contractual assumptions directly impose on those licensors and authors. All other non-permissive additional terms are considered "further restrictions" within the meaning of section 10. If the Program as you received it, or any part of it, contains a notice stating that it is governed by this License along with a term that is a further restriction, you may remove that term. If a license document contains a further restriction but permits relicensing or conveying under this License, you may add to a covered work material governed by the terms of that license document, provided that the further restriction does not survive such relicensing or conveying. If you add terms to a covered work in accord with this section, you must place, in the relevant source files, a statement of the additional terms that apply to those files, or a notice indicating where to find the applicable terms. Additional terms, permissive or non-permissive, may be stated in the form of a separately written license, or stated as exceptions; the above requirements apply either way. 8. Termination. You may not propagate or modify a covered work except as expressly provided under this License. Any attempt otherwise to propagate or modify it is void, and will automatically terminate your rights under this License (including any patent licenses granted under the third paragraph of section 11). However, if you cease all violation of this License, then your license from a particular copyright holder is reinstated (a) provisionally, unless and until the copyright holder explicitly and finally terminates your license, and (b) permanently, if the copyright holder fails to notify you of the violation by some reasonable means prior to 60 days after the cessation. Moreover, your license from a particular copyright holder is reinstated permanently if the copyright holder notifies you of the violation by some reasonable means, this is the first time you have received notice of violation of this License (for any work) from that copyright holder, and you cure the violation prior to 30 days after your receipt of the notice. Termination of your rights under this section does not terminate the licenses of parties who have received copies or rights from you under this License. If your rights have been terminated and not permanently reinstated, you do not qualify to receive new licenses for the same material under section 10. 9. Acceptance Not Required for Having Copies. You are not required to accept this License in order to receive or run a copy of the Program. Ancillary propagation of a covered work occurring solely as a consequence of using peer-to-peer transmission to receive a copy likewise does not require acceptance. However, nothing other than this License grants you permission to propagate or modify any covered work. These actions infringe copyright if you do not accept this License. Therefore, by modifying or propagating a covered work, you indicate your acceptance of this License to do so. 10. Automatic Licensing of Downstream Recipients. Each time you convey a covered work, the recipient automatically receives a license from the original licensors, to run, modify and propagate that work, subject to this License. You are not responsible for enforcing compliance by third parties with this License. An "entity transaction" is a transaction transferring control of an organization, or substantially all assets of one, or subdividing an organization, or merging organizations. If propagation of a covered work results from an entity transaction, each party to that transaction who receives a copy of the work also receives whatever licenses to the work the party's predecessor in interest had or could give under the previous paragraph, plus a right to possession of the Corresponding Source of the work from the predecessor in interest, if the predecessor has it or can get it with reasonable efforts. You may not impose any further restrictions on the exercise of the rights granted or affirmed under this License. For example, you may not impose a license fee, royalty, or other charge for exercise of rights granted under this License, and you may not initiate litigation (including a cross-claim or counterclaim in a lawsuit) alleging that any patent claim is infringed by making, using, selling, offering for sale, or importing the Program or any portion of it. 11. Patents. A "contributor" is a copyright holder who authorizes use under this License of the Program or a work on which the Program is based. The work thus licensed is called the contributor's "contributor version". A contributor's "essential patent claims" are all patent claims owned or controlled by the contributor, whether already acquired or hereafter acquired, that would be infringed by some manner, permitted by this License, of making, using, or selling its contributor version, but do not include claims that would be infringed only as a consequence of further modification of the contributor version. For purposes of this definition, "control" includes the right to grant patent sublicenses in a manner consistent with the requirements of this License. Each contributor grants you a non-exclusive, worldwide, royalty-free patent license under the contributor's essential patent claims, to make, use, sell, offer for sale, import and otherwise run, modify and propagate the contents of its contributor version. In the following three paragraphs, a "patent license" is any express agreement or commitment, however denominated, not to enforce a patent (such as an express permission to practice a patent or covenant not to sue for patent infringement). To "grant" such a patent license to a party means to make such an agreement or commitment not to enforce a patent against the party. If you convey a covered work, knowingly relying on a patent license, and the Corresponding Source of the work is not available for anyone to copy, free of charge and under the terms of this License, through a publicly available network server or other readily accessible means, then you must either (1) cause the Corresponding Source to be so available, or (2) arrange to deprive yourself of the benefit of the patent license for this particular work, or (3) arrange, in a manner consistent with the requirements of this License, to extend the patent license to downstream recipients. "Knowingly relying" means you have actual knowledge that, but for the patent license, your conveying the covered work in a country, or your recipient's use of the covered work in a country, would infringe one or more identifiable patents in that country that you have reason to believe are valid. If, pursuant to or in connection with a single transaction or arrangement, you convey, or propagate by procuring conveyance of, a covered work, and grant a patent license to some of the parties receiving the covered work authorizing them to use, propagate, modify or convey a specific copy of the covered work, then the patent license you grant is automatically extended to all recipients of the covered work and works based on it. A patent license is "discriminatory" if it does not include within the scope of its coverage, prohibits the exercise of, or is conditioned on the non-exercise of one or more of the rights that are specifically granted under this License. You may not convey a covered work if you are a party to an arrangement with a third party that is in the business of distributing software, under which you make payment to the third party based on the extent of your activity of conveying the work, and under which the third party grants, to any of the parties who would receive the covered work from you, a discriminatory patent license (a) in connection with copies of the covered work conveyed by you (or copies made from those copies), or (b) primarily for and in connection with specific products or compilations that contain the covered work, unless you entered into that arrangement, or that patent license was granted, prior to 28 March 2007. Nothing in this License shall be construed as excluding or limiting any implied license or other defenses to infringement that may otherwise be available to you under applicable patent law. 12. No Surrender of Others' Freedom. If conditions are imposed on you (whether by court order, agreement or otherwise) that contradict the conditions of this License, they do not excuse you from the conditions of this License. If you cannot convey a covered work so as to satisfy simultaneously your obligations under this License and any other pertinent obligations, then as a consequence you may not convey it at all. For example, if you agree to terms that obligate you to collect a royalty for further conveying from those to whom you convey the Program, the only way you could satisfy both those terms and this License would be to refrain entirely from conveying the Program. 13. Use with the GNU Affero General Public License. Notwithstanding any other provision of this License, you have permission to link or combine any covered work with a work licensed under version 3 of the GNU Affero General Public License into a single combined work, and to convey the resulting work. The terms of this License will continue to apply to the part which is the covered work, but the special requirements of the GNU Affero General Public License, section 13, concerning interaction through a network will apply to the combination as such. 14. Revised Versions of this License. The Free Software Foundation may publish revised and/or new versions of the GNU General Public License from time to time. Such new versions will be similar in spirit to the present version, but may differ in detail to address new problems or concerns. Each version is given a distinguishing version number. If the Program specifies that a certain numbered version of the GNU General Public License "or any later version" applies to it, you have the option of following the terms and conditions either of that numbered version or of any later version published by the Free Software Foundation. If the Program does not specify a version number of the GNU General Public License, you may choose any version ever published by the Free Software Foundation. If the Program specifies that a proxy can decide which future versions of the GNU General Public License can be used, that proxy's public statement of acceptance of a version permanently authorizes you to choose that version for the Program. Later license versions may give you additional or different permissions. However, no additional obligations are imposed on any author or copyright holder as a result of your choosing to follow a later version. 15. Disclaimer of Warranty. THERE IS NO WARRANTY FOR THE PROGRAM, TO THE EXTENT PERMITTED BY APPLICABLE LAW. EXCEPT WHEN OTHERWISE STATED IN WRITING THE COPYRIGHT HOLDERS AND/OR OTHER PARTIES PROVIDE THE PROGRAM "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESSED OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE. THE ENTIRE RISK AS TO THE QUALITY AND PERFORMANCE OF THE PROGRAM IS WITH YOU. SHOULD THE PROGRAM PROVE DEFECTIVE, YOU ASSUME THE COST OF ALL NECESSARY SERVICING, REPAIR OR CORRECTION. 16. Limitation of Liability. IN NO EVENT UNLESS REQUIRED BY APPLICABLE LAW OR AGREED TO IN WRITING WILL ANY COPYRIGHT HOLDER, OR ANY OTHER PARTY WHO MODIFIES AND/OR CONVEYS THE PROGRAM AS PERMITTED ABOVE, BE LIABLE TO YOU FOR DAMAGES, INCLUDING ANY GENERAL, SPECIAL, INCIDENTAL OR CONSEQUENTIAL DAMAGES ARISING OUT OF THE USE OR INABILITY TO USE THE PROGRAM (INCLUDING BUT NOT LIMITED TO LOSS OF DATA OR DATA BEING RENDERED INACCURATE OR LOSSES SUSTAINED BY YOU OR THIRD PARTIES OR A FAILURE OF THE PROGRAM TO OPERATE WITH ANY OTHER PROGRAMS), EVEN IF SUCH HOLDER OR OTHER PARTY HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH DAMAGES. 17. Interpretation of Sections 15 and 16. If the disclaimer of warranty and limitation of liability provided above cannot be given local legal effect according to their terms, reviewing courts shall apply local law that most closely approximates an absolute waiver of all civil liability in connection with the Program, unless a warranty or assumption of liability accompanies a copy of the Program in return for a fee. END OF TERMS AND CONDITIONS How to Apply These Terms to Your New Programs If you develop a new program, and you want it to be of the greatest possible use to the public, the best way to achieve this is to make it free software which everyone can redistribute and change under these terms. To do so, attach the following notices to the program. It is safest to attach them to the start of each source file to most effectively state the exclusion of warranty; and each file should have at least the "copyright" line and a pointer to where the full notice is found. <one line to give the program's name and a brief idea of what it does.> Copyright (C) <year> <name of author> This program is free software: you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation, either version 3 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with this program. If not, see <http://www.gnu.org/licenses/>. Also add information on how to contact you by electronic and paper mail. If the program does terminal interaction, make it output a short notice like this when it starts in an interactive mode: <program> Copyright (C) <year> <name of author> This program comes with ABSOLUTELY NO WARRANTY; for details type `show w'. This is free software, and you are welcome to redistribute it under certain conditions; type `show c' for details. The hypothetical commands `show w' and `show c' should show the appropriate parts of the General Public License. Of course, your program's commands might be different; for a GUI interface, you would use an "about box". You should also get your employer (if you work as a programmer) or school, if any, to sign a "copyright disclaimer" for the program, if necessary. For more information on this, and how to apply and follow the GNU GPL, see <http://www.gnu.org/licenses/>. The GNU General Public License does not permit incorporating your program into proprietary programs. If your program is a subroutine library, you may consider it more useful to permit linking proprietary applications with the library. If this is what you want to do, use the GNU Lesser General Public License instead of this License. But first, please read <http://www.gnu.org/philosophy/why-not-lgpl.html>.
null
[]
[]
null
null
<4.0,>=3.12
[]
[]
[]
[ "textual>=0.60", "python-dateutil>=2.8.2", "tzlocal>=5.3.1", "certifi>=2024.2.2", "packaging>=25.0", "pydantic>=2.11.7", "jinja2>=3.1.6", "click>=8.2.1", "lorem>=0.1.1", "readchar>=4.2.1", "numpy>=2.3.3", "pyperclip>=1.11.0", "tomlkit>=0.13.3", "lorem>=0.1.1; extra == \"dev\"", "pre-commit>=3.6.0; extra == \"dev\"", "pytest>=8.0.0; extra == \"test\"", "pytest-cov>=4.1.0; extra == \"test\"", "freezegun>=1.4.0; extra == \"test\"", "pytest-mock>=3.12.0; extra == \"test\"" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.14.3
2026-02-20T21:55:14.175952
tklr_dgraham-0.0.55.tar.gz
374,946
6c/a2/9b51d1630bebfbaf847f8e6c403ab7844c49ac2c273ebf5c9923456b9e27/tklr_dgraham-0.0.55.tar.gz
source
sdist
null
false
db0044aac1d4e217ab78d4653da43f2d
a2fc0ebd2575b6d5c5a3c30bd5a0cafaac884eb867d869e50433f9e03cddf455
6ca29b51d1630bebfbaf847f8e6c403ab7844c49ac2c273ebf5c9923456b9e27
null
[ "LICENSE" ]
209
2.4
twm-faust
1.17.7
Python Stream processing.
.. XXX Need to change this image to readthedocs before release .. image:: https://raw.githubusercontent.com/robinhood/faust/8ee5e209322d9edf5bdb79b992ef986be2de4bb4/artwork/banner-alt1.png =========================== Python Stream Processing =========================== Note: This project is a fork of the original **Faust** stream processing library. |build-status| |coverage| |license| |wheel| |pyversion| |pyimp| :Version: 1.16.7 :Web: http://faust.readthedocs.io/ :Download: http://pypi.org/project/faust :Source: http://github.com/robinhood/faust :Keywords: distributed, stream, async, processing, data, queue, state management .. sourcecode:: python # Python Streams # Forever scalable event processing & in-memory durable K/V store; # as a library w/ asyncio & static typing. import faust **Faust** is a stream processing library, porting the ideas from `Kafka Streams`_ to Python. It is used at `Robinhood`_ to build high performance distributed systems and real-time data pipelines that process billions of events every day. Faust provides both *stream processing* and *event processing*, sharing similarity with tools such as `Kafka Streams`_, `Apache Spark`_/`Storm`_/`Samza`_/`Flink`_, It does not use a DSL, it's just Python! This means you can use all your favorite Python libraries when stream processing: NumPy, PyTorch, Pandas, NLTK, Django, Flask, SQLAlchemy, ++ Faust requires Python 3.6 or later for the new `async/await`_ syntax, and variable type annotations. Here's an example processing a stream of incoming orders: .. sourcecode:: python app = faust.App('myapp', broker='kafka://localhost') # Models describe how messages are serialized: # {"account_id": "3fae-...", "amount": 3} class Order(faust.Record): account_id: str amount: int @app.agent(value_type=Order) async def order(orders): async for order in orders: # process infinite stream of orders. print(f'Order for {order.account_id}: {order.amount}') The Agent decorator defines a "stream processor" that essentially consumes from a Kafka topic and does something for every event it receives. The agent is an ``async def`` function, so can also perform other operations asynchronously, such as web requests. This system can persist state, acting like a database. Tables are named distributed key/value stores you can use as regular Python dictionaries. Tables are stored locally on each machine using a super fast embedded database written in C++, called `RocksDB`_. Tables can also store aggregate counts that are optionally "windowed" so you can keep track of "number of clicks from the last day," or "number of clicks in the last hour." for example. Like `Kafka Streams`_, we support tumbling, hopping and sliding windows of time, and old windows can be expired to stop data from filling up. For reliability we use a Kafka topic as "write-ahead-log". Whenever a key is changed we publish to the changelog. Standby nodes consume from this changelog to keep an exact replica of the data and enables instant recovery should any of the nodes fail. To the user a table is just a dictionary, but data is persisted between restarts and replicated across nodes so on failover other nodes can take over automatically. You can count page views by URL: .. sourcecode:: python # data sent to 'clicks' topic sharded by URL key. # e.g. key="http://example.com" value="1" click_topic = app.topic('clicks', key_type=str, value_type=int) # default value for missing URL will be 0 with `default=int` counts = app.Table('click_counts', default=int) @app.agent(click_topic) async def count_click(clicks): async for url, count in clicks.items(): counts[url] += count The data sent to the Kafka topic is partitioned, which means the clicks will be sharded by URL in such a way that every count for the same URL will be delivered to the same Faust worker instance. Faust supports any type of stream data: bytes, Unicode and serialized structures, but also comes with "Models" that use modern Python syntax to describe how keys and values in streams are serialized: .. sourcecode:: python # Order is a json serialized dictionary, # having these fields: class Order(faust.Record): account_id: str product_id: str price: float quantity: float = 1.0 orders_topic = app.topic('orders', key_type=str, value_type=Order) @app.agent(orders_topic) async def process_order(orders): async for order in orders: # process each order using regular Python total_price = order.price * order.quantity await send_order_received_email(order.account_id, order) Faust is statically typed, using the ``mypy`` type checker, so you can take advantage of static types when writing applications. The Faust source code is small, well organized, and serves as a good resource for learning the implementation of `Kafka Streams`_. **Learn more about Faust in the** `introduction`_ **introduction page** to read more about Faust, system requirements, installation instructions, community resources, and more. **or go directly to the** `quickstart`_ **tutorial** to see Faust in action by programming a streaming application. **then explore the** `User Guide`_ for in-depth information organized by topic. .. _`Robinhood`: http://robinhood.com .. _`async/await`: https://medium.freecodecamp.org/a-guide-to-asynchronous-programming-in-python-with-asyncio-232e2afa44f6 .. _`Kafka Streams`: https://kafka.apache.org/documentation/streams .. _`Apache Spark`: http://spark.apache.org .. _`Storm`: http://storm.apache.org .. _`Samza`: http://samza.apache.org .. _`Flink`: http://flink.apache.org .. _`RocksDB`: http://rocksdb.org .. _`introduction`: http://faust.readthedocs.io/en/latest/introduction.html .. _`quickstart`: http://faust.readthedocs.io/en/latest/playbooks/quickstart.html .. _`User Guide`: http://faust.readthedocs.io/en/latest/userguide/index.html Faust is... =========== **Simple** Faust is extremely easy to use. To get started using other stream processing solutions you have complicated hello-world projects, and infrastructure requirements. Faust only requires Kafka, the rest is just Python, so If you know Python you can already use Faust to do stream processing, and it can integrate with just about anything. Here's one of the easier applications you can make:: import faust class Greeting(faust.Record): from_name: str to_name: str app = faust.App('hello-app', broker='kafka://localhost') topic = app.topic('hello-topic', value_type=Greeting) @app.agent(topic) async def hello(greetings): async for greeting in greetings: print(f'Hello from {greeting.from_name} to {greeting.to_name}') @app.timer(interval=1.0) async def example_sender(app): await hello.send( value=Greeting(from_name='Faust', to_name='you'), ) if __name__ == '__main__': app.main() You're probably a bit intimidated by the `async` and `await` keywords, but you don't have to know how ``asyncio`` works to use Faust: just mimic the examples, and you'll be fine. The example application starts two tasks: one is processing a stream, the other is a background thread sending events to that stream. In a real-life application, your system will publish events to Kafka topics that your processors can consume from, and the background thread is only needed to feed data into our example. **Highly Available** Faust is highly available and can survive network problems and server crashes. In the case of node failure, it can automatically recover, and tables have standby nodes that will take over. **Distributed** Start more instances of your application as needed. **Fast** A single-core Faust worker instance can already process tens of thousands of events every second, and we are reasonably confident that throughput will increase once we can support a more optimized Kafka client. **Flexible** Faust is just Python, and a stream is an infinite asynchronous iterator. If you know how to use Python, you already know how to use Faust, and it works with your favorite Python libraries like Django, Flask, SQLAlchemy, NTLK, NumPy, SciPy, TensorFlow, etc. Installation ============ You can install Faust either via the Python Package Index (PyPI) or from source. To install using `pip`: .. sourcecode:: console $ pip install -U faust Bundles ------- Faust also defines a group of ``setuptools`` extensions that can be used to install Faust and the dependencies for a given feature. You can specify these in your requirements or on the ``pip`` command-line by using brackets. Separate multiple bundles using the comma: .. sourcecode:: console $ pip install "faust[rocksdb]" $ pip install "faust[rocksdb,uvloop,fast,redis]" The following bundles are available: Stores ~~~~~~ :``faust[rocksdb]``: for using `RocksDB`_ for storing Faust table state. **Recommended in production.** Caching ~~~~~~~ :``faust[redis]``: for using `Redis_` as a simple caching backend (Memcached-style). Codecs ~~~~~~ :``faust[yaml]``: for using YAML and the ``PyYAML`` library in streams. Optimization ~~~~~~~~~~~~ :``faust[fast]``: for installing all the available C speedup extensions to Faust core. Sensors ~~~~~~~ :``faust[datadog]``: for using the Datadog Faust monitor. :``faust[statsd]``: for using the Statsd Faust monitor. Event Loops ~~~~~~~~~~~ :``faust[uvloop]``: for using Faust with ``uvloop``. :``faust[eventlet]``: for using Faust with ``eventlet`` Debugging ~~~~~~~~~ :``faust[debug]``: for using ``aiomonitor`` to connect and debug a running Faust worker. :``faust[setproctitle]``: when the ``setproctitle`` module is installed the Faust worker will use it to set a nicer process name in ``ps``/``top`` listings. Also installed with the ``fast`` and ``debug`` bundles. Downloading and installing from source -------------------------------------- Download the latest version of Faust from http://pypi.org/project/faust You can install it by doing: .. sourcecode:: console $ tar xvfz faust-0.0.0.tar.gz $ cd faust-0.0.0 $ python setup.py build # python setup.py install The last command must be executed as a privileged user if you are not currently using a virtualenv. Using the development version ----------------------------- With pip ~~~~~~~~ You can install the latest snapshot of Faust using the following ``pip`` command: .. sourcecode:: console $ pip install https://github.com/robinhood/faust/zipball/master#egg=faust FAQ === Can I use Faust with Django/Flask/etc.? --------------------------------------- Yes! Use ``eventlet`` as a bridge to integrate with ``asyncio``. Using ``eventlet`` ~~~~~~~~~~~~~~~~~~~~~~ This approach works with any blocking Python library that can work with ``eventlet``. Using ``eventlet`` requires you to install the ``aioeventlet`` module, and you can install this as a bundle along with Faust: .. sourcecode:: console $ pip install -U faust[eventlet] Then to actually use eventlet as the event loop you have to either use the ``-L <faust --loop>`` argument to the ``faust`` program: .. sourcecode:: console $ faust -L eventlet -A myproj worker -l info or add ``import mode.loop.eventlet`` at the top of your entry point script: .. sourcecode:: python #!/usr/bin/env python3 import mode.loop.eventlet # noqa .. warning:: It's very important this is at the very top of the module, and that it executes before you import libraries. Can I use Faust with Tornado? ----------------------------- Yes! Use the ``tornado.platform.asyncio`` bridge: http://www.tornadoweb.org/en/stable/asyncio.html Can I use Faust with Twisted? ----------------------------- Yes! Use the ``asyncio`` reactor implementation: https://twistedmatrix.com/documents/17.1.0/api/twisted.internet.asyncioreactor.html Will you support Python 2.7 or Python 3.5? ------------------------------------------ No. Faust requires Python 3.6 or later, since it heavily uses features that were introduced in Python 3.6 (`async`, `await`, variable type annotations). I get a maximum number of open files exceeded error by RocksDB when running a Faust app locally. How can I fix this? -------------------------------------------------------------------------------------------------------------------- You may need to increase the limit for the maximum number of open files. The following post explains how to do so on OS X: https://blog.dekstroza.io/ulimit-shenanigans-on-osx-el-capitan/ What kafka versions faust supports? --------------------------------------- Faust supports kafka with version >= 0.10. Getting Help ============ Slack ----- For discussions about the usage, development, and future of Faust, please join the fauststream Slack. * https://fauststream.slack.com * Sign-up: https://join.slack.com/t/fauststream/shared_invite/enQtNDEzMTIyMTUyNzU2LTIyMjNjY2M2YzA2OWFhMDlmMzVkODk3YTBlYThlYmZiNTUwZDJlYWZiZTdkN2Q4ZGU4NWM4YWMyNTM5MGQ5OTg Resources ========= Bug tracker ----------- If you have any suggestions, bug reports, or annoyances please report them to our issue tracker at https://github.com/robinhood/faust/issues/ License ======= This software is licensed under the `New BSD License`. See the ``LICENSE`` file in the top distribution directory for the full license text. .. # vim: syntax=rst expandtab tabstop=4 shiftwidth=4 shiftround Contributing ============ Development of `Faust` happens at GitHub: https://github.com/robinhood/faust You're highly encouraged to participate in the development of `Faust`. Be sure to also read the `Contributing to Faust`_ section in the documentation. .. _`Contributing to Faust`: http://faust.readthedocs.io/en/latest/contributing.html Code of Conduct =============== Everyone interacting in the project's code bases, issue trackers, chat rooms, and mailing lists is expected to follow the Faust Code of Conduct. As contributors and maintainers of these projects, and in the interest of fostering an open and welcoming community, we pledge to respect all people who contribute through reporting issues, posting feature requests, updating documentation, submitting pull requests or patches, and other activities. We are committed to making participation in these projects a harassment-free experience for everyone, regardless of level of experience, gender, gender identity and expression, sexual orientation, disability, personal appearance, body size, race, ethnicity, age, religion, or nationality. Examples of unacceptable behavior by participants include: * The use of sexualized language or imagery * Personal attacks * Trolling or insulting/derogatory comments * Public or private harassment * Publishing other's private information, such as physical or electronic addresses, without explicit permission * Other unethical or unprofessional conduct. Project maintainers have the right and responsibility to remove, edit, or reject comments, commits, code, wiki edits, issues, and other contributions that are not aligned to this Code of Conduct. By adopting this Code of Conduct, project maintainers commit themselves to fairly and consistently applying these principles to every aspect of managing this project. Project maintainers who do not follow or enforce the Code of Conduct may be permanently removed from the project team. This code of conduct applies both within project spaces and in public spaces when an individual is representing the project or its community. Instances of abusive, harassing, or otherwise unacceptable behavior may be reported by opening an issue or contacting one or more of the project maintainers. This Code of Conduct is adapted from the Contributor Covenant, version 1.2.0 available at http://contributor-covenant.org/version/1/2/0/. .. |build-status| image:: https://secure.travis-ci.org/robinhood/faust.png?branch=master :alt: Build status :target: https://travis-ci.org/robinhood/faust .. |coverage| image:: https://codecov.io/github/robinhood/faust/coverage.svg?branch=master :target: https://codecov.io/github/robinhood/faust?branch=master .. |license| image:: https://img.shields.io/pypi/l/faust.svg :alt: BSD License :target: https://opensource.org/licenses/BSD-3-Clause .. |wheel| image:: https://img.shields.io/pypi/wheel/faust.svg :alt: faust can be installed via wheel :target: http://pypi.org/project/faust/ .. |pyversion| image:: https://img.shields.io/pypi/pyversions/faust.svg :alt: Supported Python versions. :target: http://pypi.org/project/faust/ .. |pyimp| image:: https://img.shields.io/pypi/implementation/faust.svg :alt: Support Python implementations. :target: http://pypi.org/project/faust/
text/x-rst
Robinhood Markets, Inc.
contact@fauststream.com
null
null
BSD 3-Clause
stream, processing, asyncio, distributed, queue, kafka
[ "Framework :: AsyncIO", "Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Natural Language :: English", "License :: OSI Approved :: BSD License", "Programming Language :: Python", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: Implementation :: CPython", "Operating System :: POSIX", "Operating System :: POSIX :: Linux", "Operating System :: MacOS :: MacOS X", "Operating System :: POSIX :: BSD", "Operating System :: Microsoft :: Windows", "Topic :: System :: Networking", "Topic :: System :: Distributed Computing" ]
[ "any" ]
http://faust.readthedocs.io/
null
>=3.10
[]
[]
[]
[ "aiohttp<4.0,>=3.8.0", "aiohttp_cors<2.0,>=0.7", "click<9.0,>=7.0", "colorclass<3.0,>=2.2", "mode-streaming>=0.4.1", "opentracing<2.0.0,>=1.3.0", "terminaltables<4.0,>=3.1", "venusian>=1.1", "yarl<3.0,>=1.0", "croniter>=0.3.16", "mypy_extensions", "intervaltree==3.1.0", "psutil>=5.9.5", "aiokafka>=0.12.0", "sortedcontainers>=2.0.0", "charset-normalizer>=3.0; extra == \"cchardet\"", "aiodns>=1.1; extra == \"fast\"", "charset-normalizer>=3.0; extra == \"fast\"", "ciso8601; extra == \"fast\"", "cython>=3.0; extra == \"fast\"", "orjson>=3.0; extra == \"fast\"", "setproctitle>=1.1; extra == \"fast\"", "python-rocksdb>=0.6.7; extra == \"rocksdb\"", "ciso8601; extra == \"ciso8601\"", "statsd~=3.3.0; extra == \"statsd\"", "datadog; extra == \"datadog\"", "rocksdict<4.0,>=0.3.25; extra == \"rocksdict\"", "redis>=4.0; extra == \"redis\"", "aiodns>=1.1; extra == \"aiodns\"", "orjson>=3.0; extra == \"orjson\"", "aiomonitor>=0.4.4; extra == \"aiomonitor\"", "pyyaml>=5.1; extra == \"yaml\"", "uvloop>=0.8.1; extra == \"uvloop\"", "setproctitle>=1.1; extra == \"setproctitle\"", "prometheus-client<1.0,>=0.20; extra == \"prometheus\"", "setproctitle>=1.1; extra == \"debug\"", "aiomonitor>=0.4.4; extra == \"debug\"", "cython>=3.0; extra == \"cython\"", "confluent-kafka>=2.0; extra == \"ckafka\"" ]
[]
[]
[]
[ "Bug Reports, https://github.com/robinhood/faust/issues", "Source, https://github.com/robinhood/faust", "Documentation, https://faust.readthedocs.io/" ]
twine/6.2.0 CPython/3.12.2
2026-02-20T21:55:05.529224
twm_faust-1.17.7.tar.gz
680,960
fb/8c/2a7d49f73970e1e738f36d10143eb430a5e44a237b34dba8ece78dce45cb/twm_faust-1.17.7.tar.gz
source
sdist
null
false
332ab7e883e4268d9ab1a2818e88a9c7
35802d32e62aebf209849002689ef498267a7887a424a10fe19ae3b3b9684fc7
fb8c2a7d49f73970e1e738f36d10143eb430a5e44a237b34dba8ece78dce45cb
null
[ "LICENSE" ]
226
2.4
astar-utils
0.5.1
Contains commonly-used utilities for AstarVienna's projects.
# Astar Utils [![Tests](https://github.com/AstarVienna/astar-utils/actions/workflows/tests.yml/badge.svg)](https://github.com/AstarVienna/astar-utils/actions/workflows/tests.yml) [![Poetry](https://img.shields.io/endpoint?url=https://python-poetry.org/badge/v0.json)](https://python-poetry.org/) ![dev version](https://img.shields.io/badge/dynamic/toml?url=https%3A%2F%2Fraw.githubusercontent.com%2FAstarVienna%2Fastar-utils%2Fmain%2Fpyproject.toml&query=%24.project.version&label=dev%20version&color=teal) [![codecov](https://codecov.io/gh/AstarVienna/astar-utils/graph/badge.svg)](https://codecov.io/gh/AstarVienna/astar-utils) [![PyPI - Version](https://img.shields.io/pypi/v/astar-utils)](https://pypi.org/project/astar-utils/) ![PyPI - Python Version](https://img.shields.io/pypi/pyversions/astar-utils) ![GitHub Release Date](https://img.shields.io/github/release-date/AstarVienna/astar-utils) [![License: GPL v3](https://img.shields.io/badge/License-GPLv3-blue.svg)](https://www.gnu.org/licenses/gpl-3.0) This package is devloped and maintained by [Astar Vienna](https://github.com/AstarVienna) and contains commonly-used utilities for the group's projects to avoid both duplicating code and circular dependencies. ## Contents The package currently contains the following public functions and classes: - `NestedMapping`: a `dict`-like structure supporting !-style nested keys. - `RecursiveNestedMapping`: a subclass of `NestedMapping` also supporting keys that reference other !-style keys. - `NestedChainMap`: a subclass of `collections.ChainMap` supporting instances of `RecursiveNestedMapping` as levels and referencing !-style keys across chain map levels. - `is_bangkey()`: simple convenience function to check if something is a !-style key. - `is_nested_mapping()`: convenience function to check if something is a mapping containing a least one other mapping as a value. - `UniqueList`: a `list`-like structure with no duplicate elements and some convenient methods. - `Badge` and subclasses: a family of custom markdown report badges. See docstring for details. - `BadgeReport`: context manager for collection and generation of report badges. See docstring for details and usage. - `get_logger()`: convenience function to get (or create) a logger with given `name` as a child of the universal `astar` logger. - `get_astar_logger()`: convenience function to get (or create) a logger with the name `astar`, which serves as the root for all A*V packages and applications. - `SpectralType`: a class to parse, store and compare spectral type designations. ### Loggers module - `loggers.ColoredFormatter`: a subclass of `logging.Formatter` to produce colored logging messages for console output. ## Dependencies Dependencies are intentionally kept to a minimum for simplicity. Current dependencies are: - `more-itertools` - `pyyaml` Version requirement for these dependencies can be found in the `pyproject.toml` file.
text/markdown
Fabian Haberhauer
fabian.haberhauer@univie.ac.at
Fabian Haberhauer
fabian.haberhauer@univie.ac.at
GPL-3.0-or-later
null
[ "Intended Audience :: Science/Research", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Topic :: Scientific/Engineering :: Astronomy", "Topic :: Utilities" ]
[]
null
null
<3.15,>=3.10
[]
[]
[]
[ "colorama<1.0,>=0.4.6", "more-itertools<11.0.0,>=10.2.0", "pyyaml<7.0.0,>=6.0.3" ]
[]
[]
[]
[ "Changelog, https://github.com/AstarVienna/astar-utils/releases", "Repository, https://github.com/AstarVienna/astar-utils" ]
poetry/2.3.0 CPython/3.12.3 Linux/6.11.0-1018-azure
2026-02-20T21:54:38.599248
astar_utils-0.5.1-py3-none-any.whl
29,114
db/f8/a1b0e104bf0e116966e663213c5e98aa5b1b92e57ef272f7fa70ee4f2d22/astar_utils-0.5.1-py3-none-any.whl
py3
bdist_wheel
null
false
79384bdd8d430fa2d71a6f87ef9b4809
295f520f24a636e92cdc916a9adc36724f7d2091283324881a2d18e083572c17
dbf8a1b0e104bf0e116966e663213c5e98aa5b1b92e57ef272f7fa70ee4f2d22
null
[ "LICENSE" ]
340
2.4
antioch-py
3.1.2
Antioch Python Module SDK
# antioch-py Python SDK for the [Antioch](https://antioch.com) autonomy simulation platform. ## Overview The antioch-py package provides two components: ### Module SDK (`antioch.module`) The Module SDK is a framework for building Antioch modules in Python. Modules are containerized components that run alongside your simulation, processing sensor data and producing outputs. Each module runs in its own Docker container and communicates with the simulation through the Antioch runtime. Install the SDK in your module's Dockerfile to read sensors, run inference, and publish results. ```python from antioch.module import Execution, Module def process_radar(execution: Execution) -> None: scan = execution.read_radar("sensor") if scan is not None and len(scan.detections) > 0: execution.output("detections").set(scan) if __name__ == "__main__": module = Module() module.register("radar_node", process_radar) module.spin() ``` ### Session SDK (`antioch.session`) The Session SDK is a client library for orchestrating Antioch simulations. Use it from Python scripts or Jupyter notebooks to programmatically build scenes, load assets, spawn robots, control simulation playback, and record data. The Session SDK connects to your Antioch deployment and provides a high-level API for automation and experimentation. ```python from antioch.session import Scene, Session, Task, TaskOutcome session = Session() scene = Scene() # Load environment and robot scene.add_asset(path="/World/environment", name="warehouse", version="1.0.0") ark = scene.add_ark(name="my_robot", version="0.1.0") # Run simulation task = Task() task.start(mcap_path="/tmp/recording.mcap") scene.step(1_000_000) # step 1 second task.finish(outcome=TaskOutcome.SUCCESS) ``` ## Installation To install in your Python environment: ```bash pip install antioch-py ``` To install in your Python-based Docker image (e.g. for an Antioch module): ```dockerfile FROM python:3.12-slim RUN pip install antioch-py COPY . /app WORKDIR /app CMD ["python", "module.py"] ``` ## Documentation Visit [antioch.com](https://antioch.com) for full documentation. ## License MIT
text/markdown
null
Antioch Robotics <support@antioch.dev>
null
null
null
robotics, simulation, middleware, sdk, modules
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.12", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules" ]
[]
null
null
<3.13,>=3.10
[]
[]
[]
[ "click>=8.0.0", "docker>=7.0.0", "eclipse-zenoh>=1.5.0", "foxglove-sdk>=0.14.1", "httpx>=0.27.0", "loguru>=0.7.3", "msgpack==1.1.1", "msgpack>=1.1.1", "numpy==1.26.0", "ormsgpack>=1.6.0", "pydantic>=2.11.6", "pydantic>=2.11.7", "pyyaml>=6.0.2", "scipy==1.15.3", "sortedcontainers-stubs>=2.4.3", "sortedcontainers>=2.4.0" ]
[]
[]
[]
[ "Homepage, https://antioch.com" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:54:24.452585
antioch_py-3.1.2.tar.gz
79,087
be/29/5ec7236e99cdc2c115cd9f32de8754b0166c46bf19d507d5eb5d37827b6a/antioch_py-3.1.2.tar.gz
source
sdist
null
false
257b91c4a4f61205a92b702dbb3b3df0
0e6aff22116045a27e34f095c043d147e3d9542cb693087bda00e1bb02e8ec21
be295ec7236e99cdc2c115cd9f32de8754b0166c46bf19d507d5eb5d37827b6a
MIT
[ "LICENSE" ]
209
2.4
agentready
2.29.1
Assess and bootstrap git repositories for AI-assisted development with automated remediation and continuous learning
# AgentReady Repository Scorer [![codecov](https://codecov.io/gh/ambient-code/agentready/branch/main/graph/badge.svg)](https://codecov.io/gh/ambient-code/agentready) [![Tests](https://github.com/ambient-code/agentready/actions/workflows/ci.yml/badge.svg)](https://github.com/ambient-code/agentready/actions/workflows/ci.yml) Assess git repositories against evidence-based attributes for AI-assisted development readiness. > **📚 Research-Based Assessment**: AgentReady's attributes are derived from [comprehensive research](RESEARCH_REPORT.md) analyzing 50+ authoritative sources including **Anthropic**, **Microsoft**, **Google**, **ArXiv**, and **IEEE/ACM**. Each attribute is backed by peer-reviewed research and industry best practices. [View full research report →](RESEARCH_REPORT.md) ## Overview AgentReady evaluates your repository across multiple dimensions of code quality, documentation, testing, and infrastructure to determine how well-suited it is for AI-assisted development workflows. The tool generates comprehensive reports with: - **Overall Score & Certification**: Platinum/Gold/Silver/Bronze based on comprehensive attribute assessment - **Interactive HTML Reports**: Filter, sort, and explore findings with embedded guidance - **Version-Control-Friendly Markdown**: Track progress over time with git-diffable reports - **Actionable Remediation**: Specific tools, commands, and examples to improve each attribute - **Schema Versioning**: Backwards-compatible report format with validation and migration tools ## Quick Start ### Container (Recommended) ```bash # Login to GitHub Container Registry (required for private image) podman login ghcr.io # Pull container podman pull ghcr.io/ambient-code/agentready:latest # Create output directory mkdir -p ~/agentready-reports # Assess AgentReady itself git clone https://github.com/ambient-code/agentready /tmp/agentready podman run --rm \ -v /tmp/agentready:/repo:ro \ -v ~/agentready-reports:/reports \ ghcr.io/ambient-code/agentready:latest \ assess /repo --output-dir /reports # Assess your repository # For large repos, add -i flag to confirm the size warning podman run --rm \ -v /path/to/your/repo:/repo:ro \ -v ~/agentready-reports:/reports \ ghcr.io/ambient-code/agentready:latest \ assess /repo --output-dir /reports # Open reports open ~/agentready-reports/report-latest.html ``` [See full container documentation →](CONTAINER.md) ### Python Package ```bash # Install pip install agentready # Assess AgentReady itself git clone https://github.com/ambient-code/agentready /tmp/agentready agentready assess /tmp/agentready # Create virtual environment python3 -m venv .venv source .venv/bin/activate # On Windows: .venv\Scripts\activate # Install dependencies pip install -e ".[dev]" ``` ### Run Directly via uv (Optional, No Install Required) If you use **uv**, you can run AgentReady directly from GitHub without cloning or installing: ```bash uvx --from git+https://github.com/ambient-code/agentready agentready -- assess . ``` To install it as a reusable global tool: ```bash uv tool install --from git+https://github.com/ambient-code/agentready agentready ``` After installing globally: ```bash agentready assess . ``` ### Harbor CLI (for Benchmarks) Harbor is required for running Terminal-Bench evaluations: ```bash # AgentReady will prompt to install automatically, or install manually: uv tool install harbor # Alternative: Use pip if uv is not available pip install harbor # Verify installation harbor --version ``` **Skip automatic checks**: If you prefer to skip the automatic Harbor check (for advanced users): ```bash agentready benchmark --skip-preflight --subset smoketest ``` ### Assessment Only For one-time analysis without infrastructure changes: ```bash # Assess current repository agentready assess . # Assess another repository agentready assess /path/to/your/repo # Specify custom configuration agentready assess /path/to/repo --config my-config.yaml # Custom output directory agentready assess /path/to/repo --output-dir ./reports ``` ### Example Output ``` Assessing repository: myproject Repository: /Users/username/myproject Languages detected: Python (42 files), JavaScript (18 files) Evaluating attributes... [████████████████████████░░░░░░░░] 23/25 (2 skipped) Overall Score: 72.5/100 (Silver) Attributes Assessed: 23/25 Duration: 2m 7s Reports generated: HTML: .agentready/report-latest.html Markdown: .agentready/report-latest.md ``` ## Features ### Evidence-Based Attributes Evaluated across 13 categories: 1. **Context Window Optimization**: CLAUDE.md files, concise docs, file size limits 2. **Documentation Standards**: README structure, inline docs, ADRs 3. **Code Quality**: Cyclomatic complexity, file length, type annotations, code smells 4. **Repository Structure**: Standard layouts, separation of concerns 5. **Testing & CI/CD**: Coverage, test naming, pre-commit hooks 6. **Dependency Management**: Lock files, freshness, security 7. **Git & Version Control**: Conventional commits, gitignore, templates 8. **Build & Development**: One-command setup, dev docs, containers 9. **Error Handling**: Clear messages, structured logging 10. **API Documentation**: OpenAPI/Swagger specs 11. **Modularity**: DRY principle, naming conventions 12. **CI/CD Integration**: Pipeline visibility, branch protection 13. **Security**: Scanning automation, secrets management ### Tier-Based Scoring Attributes are weighted by importance: - **Tier 1 (Essential)**: 50% of total score - CLAUDE.md, README, types, layouts, lock files - **Tier 2 (Critical)**: 30% of total score - Tests, commits, build setup - **Tier 3 (Important)**: 15% of total score - Complexity, logging, API docs - **Tier 4 (Advanced)**: 5% of total score - Security scanning, performance benchmarks Missing essential attributes (especially CLAUDE.md at 10% weight) has 10x the impact of missing advanced features. ### Interactive HTML Reports - Filter by status (Pass/Fail/Skipped) - Sort by score, tier, or category - Search attributes by name - Collapsible sections with detailed evidence - Color-coded score indicators - Certification ladder visualization - Works offline (no CDN dependencies) ### Customization Create `.agentready-config.yaml` to customize weights: ```yaml weights: claude_md_file: 0.15 # Increase importance (default: 0.10) test_coverage: 0.05 # Increase importance (default: 0.03) conventional_commits: 0.01 # Decrease importance (default: 0.03) # Other attributes use defaults, rescaled to sum to 1.0 excluded_attributes: - performance_benchmarks # Skip this attribute output_dir: ./custom-reports ``` ## CLI Reference ```bash # Assessment commands agentready assess PATH # Assess repository at PATH agentready assess PATH --verbose # Show detailed progress agentready assess PATH --config FILE # Use custom configuration agentready assess PATH --output-dir DIR # Custom report location # Configuration commands agentready --validate-config FILE # Validate configuration agentready --generate-config # Create example config # Research report management agentready research-version # Show bundled research version agentready research validate FILE # Validate research report agentready research init # Generate new research report agentready research add-attribute FILE # Add attribute to report agentready research bump-version FILE # Update version agentready research format FILE # Format research report # Utility commands agentready --version # Show tool version agentready --help # Show help message ``` ## Architecture AgentReady follows a library-first design: - **Models**: Data entities (Repository, Assessment, Finding, Attribute) - **Assessors**: Independent evaluators for each attribute category - **Services**: Scanner (orchestration), Scorer (calculation), LanguageDetector - **Reporters**: HTML and Markdown report generators - **CLI**: Thin wrapper orchestrating assessment workflow ## Development ### Run Tests ```bash # Run all tests with coverage pytest # Run specific test suite pytest tests/unit/ pytest tests/integration/ pytest tests/contract/ # Run with verbose output pytest -v -s ``` ### Code Quality ```bash # Format code black src/ tests/ # Sort imports isort src/ tests/ # Lint code flake8 src/ tests/ --ignore=E501 # Run all checks black . && isort . && flake8 . ``` ### Project Structure ``` src/agentready/ ├── cli/ # Click-based CLI entry point ├── assessors/ # Attribute evaluators (13 categories) ├── models/ # Data entities ├── services/ # Core logic (Scanner, Scorer) ├── reporters/ # HTML and Markdown generators ├── templates/ # Jinja2 HTML template └── data/ # Bundled research report and defaults tests/ ├── unit/ # Unit tests for individual components ├── integration/ # End-to-end workflow tests ├── contract/ # Schema validation tests └── fixtures/ # Test repositories ``` ## Research Foundation All attributes are derived from evidence-based research with 50+ citations from: - Anthropic (Claude Code documentation, engineering blog) - Microsoft (Code metrics, Azure DevOps best practices) - Google (SRE handbook, style guides) - ArXiv (Software engineering research papers) - IEEE/ACM (Academic publications on code quality) See `src/agentready/data/RESEARCH_REPORT.md` for complete research report. ## License MIT License - see LICENSE file for details. ## Contributing Contributions welcome! Please ensure: - All tests pass (`pytest`) - Code is formatted (`black`, `isort`) - Linting passes (`flake8`) - Test coverage >80% ## Support - Documentation: See `/docs` directory - Issues: Report at GitHub Issues - Questions: Open a discussion on GitHub --- **Quick Start**: `pip install -e ".[dev]" && agentready assess .` - Ready in <5 minutes!
text/markdown
null
Jeremy Eder <jeder@redhat.com>
null
null
MIT
null
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Software Development :: Quality Assurance" ]
[]
null
null
>=3.12
[]
[]
[]
[ "click>=8.1.0", "jinja2>=3.1.0", "pyyaml>=6.0", "gitpython>=3.1.0", "radon>=6.0.0", "lizard>=1.17.0", "anthropic>=0.74.0", "jsonschema>=4.17.0", "requests>=2.31.0", "pydantic>=2.0.0", "pandas>=2.0.0", "plotly>=5.0.0", "scipy>=1.10.0", "PyGithub>=2.1.1", "pytest>=7.4.0; extra == \"dev\"", "pytest-cov>=4.1.0; extra == \"dev\"", "black>=23.0.0; extra == \"dev\"", "isort>=5.12.0; extra == \"dev\"", "flake8>=6.0.0; extra == \"dev\"", "ruff>=0.1.0; extra == \"dev\"" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:54:13.026023
agentready-2.29.1.tar.gz
322,293
cd/d4/260fea4de0adc50bb38e45645079f4766dda73ac517c499ce7e7e39c3380/agentready-2.29.1.tar.gz
source
sdist
null
false
5c9bc14d745517e5cbdab782cb6501c5
3255f98ad053d134ebcd8db31b29d3fc68b830eca2cb7781d24c3b53382ac95d
cdd4260fea4de0adc50bb38e45645079f4766dda73ac517c499ce7e7e39c3380
null
[ "LICENSE" ]
198
2.4
sthenos
0.1.32
A Python-scriptable load testing tool (Go binary wrapper)
# Sthenos **Sthenos** is a modern, high-performance load testing tool written in Go. It combines the raw power of Go with the simplicity of Python-like scripting (Starlark) for an effortless developer experience. ## Features - **High Performance**: Built on Go, using lightweight goroutines for Virtual Users (VUs). - **Python-like Scripting**: Write test scripts in Starlark (a dialect of Python). - **CLI Driven**: Simple, powerful command-line interface. - **Metrics**: Detailed breakdown of HTTP request lifecycle (DNS, TLS, TTFB, etc.). - **Thresholds**: Define pass/fail criteria directly in your script. - **Stages**: Ramp up/down VUs to simulate real-world traffic patterns. - **Zero-Dependency Installation**: `pip install sthenos` works instantly (bundles pre-compiled binaries). - **K6 Compatibility**: Supports Stages, Checks, Thresholds, and Setup/Teardown lifecycles. - **Detailed Metrics**: Breakdown of DNS, TLS, TTFB, and more. - **Auto-JSON**: Pass `json={...}` to automatically serialize, set headers, and send. - **Query Params**: Pass `query={'q': 'foo'}` to automatically encode URL parameters. ## Installation ### Via PyPI (Recommended) ```bash pip install sthenos ``` ### From Source ```bash git clone https://github.com/richard24se/Sthenos.git cd Sthenos/sthenos go build -o sthenos main.go ``` ## Usage ### Basic Run ```bash ./sthenos run script.py ``` ### Options Override script defaults via CLI flags: ```bash ./sthenos run \ --vus 10 \ --duration 30s \ --out json=results.json \ --env HOST=api.staging.com \ script.py ``` | Flag | Description | Default | | :--- | :--- | :--- | | `--vus` | Number of Virtual Users | 1 | | `--duration` | Test duration (e.g., `10s`, `1m`) | `10s` | | `--env` | Set environment variable (`KEY=VAL`) | - | | `--out` | Output format (currently `json=path`) | - | | `--graceful-stop` | Timeout to wait for active VUs | `30s` | | `--verbose` | Show detailed timing breakdown | `false` | ## API Reference ### HTTP Module (`sthenos.http`) Sthenos supports all standard HTTP methods. **Methods:** - `http.get(url, params=?, query=?, stop_failure=?)` - `http.post(url, body=?, json=?, params=?, query=?, stop_failure=?)` - `http.put(url, body=?, json=?, params=?, query=?, stop_failure=?)` - `http.del(url, body=?, json=?, params=?, query=?, stop_failure=?)` (alias: `delete`) - `http.patch(url, body=?, json=?, params=?, query=?, stop_failure=?)` - `http.head(url, params=?, query=?, stop_failure=?)` - `http.options(url, params=?, query=?, stop_failure=?)` **Arguments:** - `url` (string): Target URL. - `body` (string): Raw request body (text/xml/binary). - `json` (dict): Auto-serializes to JSON and sets `Content-Type: application/json`. - `query` (dict): Appends URL query parameters (e.g., `?limit=10`). - `headers` (dict): Custom HTTP headers (e.g., `{'Authorization': 'Bearer ...'}`). - `params` (dict): **Deprecated**. Use `headers` instead. Backward compatibility only. - `stop_failure` (bool): Overrides the global stop-on-failure setting for this request. **Response Object:** Returned by all HTTP methods. - `res.status` (int): HTTP status code (e.g., 200, 404). - `res.body` (string): Response body text. - `res.headers` (dict): Response headers. ### Checks (`sthenos.check`) Validate responses without stopping the test. ```python check(res, { "is status 200": lambda r: r.status == 200, "body has error": lambda r: "error" not in r.body, }) ``` ### Global Objects - `ENV` (dict): Access environment variables passed via `--env key=val`. - `sleep(seconds)`: Pause VU execution. ## Examples & Features Sthenos scripting is designed to be intuitive. Below are examples covering all major features. Scripts are written in **Starlark**, which looks and feels like Python. ### 1. Basic Request (`01_basic.py`) The simplest test—hitting an endpoint repeatedly. ```python # Command: sthenos run --vus 5 --duration 5s examples/01_basic.py # 01_basic.py: Minimal load test from sthenos import http, sleep def default(): # Perform a simple GET request http.get("http://localhost:9001/") sleep(0.2) sleep(1) ``` ### 2. Traffic Stages (`02_stages.py`) Simulate ramping up and down traffic patterns. ```python # Command: sthenos run examples/02_stages.py # 02_stages.py: Ramping VUs using stages from sthenos import http, sleep # Define ramping stages options = { "stages": [ {"duration": "2s", "target": 5}, # Ramp up to 5 VUs {"duration": "5s", "target": 10}, # Stay at 5 VUs {"duration": "2s", "target": 0}, # Ramp down to 0 ] } def default(): http.get("http://localhost:9001/") sleep(0.5) ``` ### 3. Fail/Pass Criteria (`03_thresholds.py`) Ensure your system meets specific SLIs. The test exits with code 1 if thresholds fail. ```python # Command: sthenos run examples/03_thresholds.py # 03_thresholds.py: Defining Pass/Fail criteria from sthenos import http, check, sleep options = { # Thresholds allow CI/CD integration "thresholds": { "http_req_duration": ["p(95) < 500"], # 95% of requests must be faster than 500ms "checks": ["rate > 0.95"], # 95% of checks must pass "http_req_failed": ["rate < 0.01"], # Less than 1% errors }, "stages": [ {"duration": "5s", "target": 100}, ] } def default(ctx): res = http.get("http://localhost:9001/") # This check feeds the "checks" metric check(res, { "is status 200": lambda r: r.status == 200, }) sleep(1) ``` ### 4. Validating Responses (`04_env_vars.py`) This example shows how to use environment variables. ```python # Command: sthenos run --env HOST=https://httpbin.org --env USER=admin examples/04_env_vars.py # 04_env_vars.py: Using Environment Variables from sthenos import http, check, sleep, ENV def setup(): # Access injected env vars via global ENV dict host = ENV.get("HOST", "http://localhost:9001") user = ENV.get("USER", "guest") print("[Setup] Target: {host}, User: {user}".format(host=host, user=user)) return {"host": host} def default(ctx): url = "{host}/".format(host=ctx["host"]) http.get(url) sleep(1) ``` ### 5. POST JSON Data (`05_post_json.py`) Send JSON payloads easily. ```python # Command: sthenos run --vus 2 --duration 5s examples/05_post_json.py # 05_post_json.py: Sending POST requests from sthenos import http, check, sleep def default(ctx): # TODO: Current MVP implementation supports basic POST # Future: support full JSON body as dict payload = '{"foo": "bar"}' res = http.post("http://localhost:9001/post", payload) check(res, { "status is 200": lambda r: r.status == 200, }) sleep(0.1) ``` ### 6. Environment Variables (`06_checks.py`) This example shows how to use checks to validate responses. ```python # Command: sthenos run --vus 5 --duration 5s examples/06_checks.py # 06_checks.py: Validating responses from sthenos import http, check, sleep def default(ctx): res = http.get("http://localhost:9001/get") # Checks don't stop execution but are recorded success = check(res, { "status is 200": lambda r: r.status == 200, "body size > 0": lambda r: len(r.body) > 0, }) if not success: print("Request failed validation!") sleep(1) ``` ### 7. Full Lifecycle (`07_lifecycle.py`) Use `setup()` (once globally) to prepare data, and `teardown()` (once globally) to clean up. ```python # Command: sthenos run examples/07_lifecycle.py # 07_lifecycle.py: Demonstrating the full test lifecycle # Init -> Setup(once) -> VU(many) -> Teardown(once) from sthenos import http, check, sleep # 1. INIT PHASE # Executed once per VU process/thread print("[Phase] Init") # 2. SETUP PHASE # Executed once globally. Returns 'ctx' data. def setup(): print("[Phase] Setup: Generating Test Data...") return {"token": "secret_abc_123", "items": [1, 2, 3]} # 3. VU PHASE # Executed repeatedly. Receives 'ctx'. def default(ctx): # print("[Phase] VU: Using token {}".format(token)) # Uncomment to see per-VU logs # Simulate work res = http.get("http://localhost:9001/headers", {"headers": {"Authorization": ctx["token"]}}) # Validate that server received the token via Response Headers check( res, { "status is 200": lambda r: r.status == 200, }, ) check( res, { "token verified": lambda r: r.headers.get("X-Token-Received") == ctx["token"], }, ) sleep(0.5) # 4. TEARDOWN PHASE # Executed once globally. Receives 'ctx'. def teardown(ctx): print("[Phase] Teardown: Data {} clean up complete.".format(ctx["items"])) ``` ### 8. Complete Scenario (`08_ecommerce.py`) Combine everything to simulate a real user journey. ```python # Command: sthenos run examples/08_ecommerce.py # 08_ecommerce.py: User Journey Simulation # Features: Stages, Thresholds, Env Vars, Setup, Checks, Sleep, POST from sthenos import http, check, sleep, ENV # 1. Option Configuration options = { # Ramping Load Pattern "stages": [ {"duration": "10s", "target": 50}, # Ramp up to 5 users {"duration": "5s", "target": 5}, # Stay at 5 users {"duration": "2s", "target": 0}, # Ramp down to 0 ], # Pass/Fail Criteria "thresholds": { "http_req_duration": ["p(95)<500"], # 95% of requests must be faster than 500ms "checks": ["rate>0.95"], # 95% of checks must pass "http_req_failed": ["rate<0.01"], # Less than 1% failure }, } # 2. Setup Phase def setup(): # Use Environment Variable for dynamic targeting # usage: sthenos run --env HOST=https://my-api.com examples/08_ecommerce.py # usage: sthenos run --env HOST=https://my-api.com examples/08_ecommerce.py host = ENV.get("HOST", "http://localhost:9001") print("Setup: Targeting {}".format(host)) return {"host": host} # 3. VU Logic def default(ctx): host = ctx["host"] # Step 1: Browse Home Page res = http.get("{}/get?page=home".format(host)) check(res, { "home status 200": lambda r: r.status == 200, }) sleep(1.0) # User thinks # Step 2: View Product res = http.get("{}/get?product=123".format(host)) check(res, { "product status 200": lambda r: r.status == 200, }) sleep(0.5) # Step 3: Add to Cart (POST) # Note: Sthenos supports string body for POST payload = '{"product_id": 123, "qty": 1}' params = { "headers": { "Content-Type": "application/json", "Authorization": "Bearer secret_token", } } res = http.post("{}/post".format(host), payload, params) check(res, { "cart add status 200": lambda r: r.status == 200, }) sleep(1.0) def teardown(data): print("Teardown: Test Complete.") ``` ### 9. Failure & Thresholds (`09_failure.py`) Demonstrates how failures are reported when defining thresholds that expect 200 OK responses but encounter 404s. ```python # Command: sthenos run examples/09_failure.py # 09_failure.py: Demonstrating Failures & Thresholds # This script is INTENDED TO FAIL to show how Sthenos reports errors. from sthenos import http, check, sleep options = { # Define thresholds that will definitely fail "thresholds": { "checks": ["rate>0.99"], # Expect 99% pass rate (will fail) "http_req_duration": ["p(95)<10"], # Expect extremely fast <10ms (will fail) }, } def default(ctx): # Call a URL that returns 404 (Fail) res = http.get("http://localhost:9001/status/404") check(res, { "is status 200": lambda r: r.status == 200, }) sleep(1) ``` ### 10. Loops & Iteration (`10_loops.py`) Demonstrates iterating over lists and ranges in Starlark. ```python # Command: sthenos run examples/10_loops.py # 10_loops.py: Loops & Iteration from sthenos import http, sleep def default(ctx): items = ["item_a", "item_b", "item_c"] for item in items: url = "http://localhost:9001/get?item=" + item http.get(url) sleep(0.1) for i in range(3): url = "http://localhost:9001/get?id=" + str(i) http.get(url) sleep(0.1) ``` ### 11. Testing Response Output and HTTP Args (`11_http_args.py`) Validating parsed JSON in responses and dynamically passing query parameters. ```python # Command: sthenos run examples/11_http_args.py # 11_http_args.py from sthenos import http, check, sleep def default(ctx): res = http.post( "http://localhost:9001/post", json={"num": 123, "bool": True, "list": [1, 2, 3]} ) check(res, {"json_status": lambda r: r.status == 200}) res_query = http.get( "http://localhost:9001/get", query={"page": "1", "product": "sthenos_v1"} ) check(res_query, {"query_status": lambda r: r.status == 200}) sleep(1) ``` ### 12. POST JSON Auto-Serialization (`12_http_json.py`) Automatically serialize dictionaries to JSON and set Content-Type. ```python # Command: sthenos run examples/12_http_json.py # 12_http_json.py from sthenos import http, check, sleep def default(ctx): # Pass dict to 'json' argument -> converts to body + sets Content-Type res = http.post( "http://localhost:9001/post", json={"tool": "sthenos", "active": True} ) check(res, { "status is 200": lambda r: r.status == 200, }) sleep(1) ``` ### 13. Raw HTTP Body (`13_http_body.py`) Send raw text, XML, or binary data with custom headers. ```python # Command: sthenos run examples/13_http_body.py # 13_http_body.py from sthenos import http, check, sleep def default(ctx): # XML Post body_xml = "<user><name>Fausto</name></user>" res = http.post( "http://localhost:9001/echo_xml", body=body_xml, params={"headers": {"Content-Type": "application/xml"}} ) check(res, {"status is 200": lambda r: r.status == 200}) sleep(1) ``` ### 14. Iteration Failures (`14_iteration_failed.py`) Demonstrates how network errors or timeouts trigger an efficient "Failed" status without stopping the test (unless configured). ```python # Command: sthenos run examples/14_iteration_failed.py # 14_iteration_failed.py from sthenos import http, sleep def default(): # Successful request http.get("http://localhost:9001/") # Failed request (Connection Refused) # By default, this aborts the iteration immediately (stop_failure=True) http.get("http://localhost:12345/") # This line is never reached by default sleep(1) ``` ### 15. Stop-on-Failure Configuration (`15_iteration_failed_options.py`) Control whether a failure keeps the VU alive (continue) or aborts the iteration. ```python # Command: sthenos run examples/15_iteration_failed_options.py # 15_iteration_failed_options.py from sthenos import http options = { "stop_failure": False, # Global Override: Don't stop on error } def default(): # Will fail, but continue because of global option http.get("http://localhost:12345/") # Explicit override: This WILL stop the iteration http.get("http://localhost:12345/", stop_failure=True) ``` ### 16. All HTTP Methods (`16_all_methods.py`) Demonstrating all standard HTTP methods supported by Sthenos. ```python # Command: sthenos run examples/16_all_methods.py # 16_all_methods.py from sthenos import http, check, sleep def default(ctx): url = "http://localhost:9001/all-methods" res_get = http.get(url) res_post = http.post(url, json={"hello": "world"}) res_put = http.put(url, json={"updated": True}) res_patch = http.patch(url, json={"partial": True}) try: # 'del' is a reserved keyword in Python Starlark, so use dict/getattr workaround if needed # Or you can map it in engine. res_del = getattr(http, "del")(url) except Exception: pass res_head = http.head(url) res_options = http.options(url) sleep(1) ``` ## Metrics Output Sthenos produces a k6-inspired summary to the console: ```text _____ _______ _ _ ______ _ _ ____ _____ / ____|__ __| | | | ____| \ | |/ __ \ / ____| | (___ | | | |__| | |__ | \| | | | | (___ \___ \ | | | __ | __| | . | | | |\___ \ ____) | | | | | | | |____| |\ | |__| |____) | |_____/ |_| |_| |_|______|_| \_|\____/|_____/ Initial VUs: 100, Duration: 5s Running Setup... Setup: Targeting http://localhost:9001 Mode: Forced Constant (CLI Overrides). VUs: 100, Duration: 5s Starting Constant Load: 100 VUs for 5s running [=========================] 5s / 5s Stopping VUs and waiting for graceful exit... Running Teardown... Teardown: Test Complete. ✓ checks...........................: 100.00% ✓ 600 / 600 data_received......................: 118.4 kB 23.2 kB/s data_sent..........................: 49.4 kB 9.7 kB/s http_req_duration..................: avg=10.67ms min=0.00ms med=1.06ms max=85.21ms p(90)=44.67ms p(95)=65.27ms http_reqs..........................: 600 117.47/s http_req_failed....................: 0 0.00% vus................................: 100 min=100 max=100 vus_max............................: 100 min=100 max=100 --- Thresholds --- http_req_duration p(95)<500 ... PASS checks rate>0.95 ... PASS http_req_failed rate<0.01 ... PASS ------------------ ``` ## Comparisons | Feature | k6 (OSS) | Sthenos | | :--- | :--- | :--- | | **Language** | Go | Go | | **Scripting** | JavaScript (ES6) | Starlark (Python dialect) | | **Engine** | goja | go.starlark.net | | **Concurrency** | Goroutines | Goroutines | | **Ecosystem** | Mature | Experimental | ## License STHENOS LICENSE
text/markdown
null
Aldo Richard Santillán Echevarría <richard.24.se@gmail.com>
null
null
STHENOS LICENSE Version 1.1 Copyright (c) 2025 Richard Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to use, copy, modify, and distribute the Software for both personal and commercial purposes, subject to the following conditions: 1. Attribution The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. You must inform users of the software of its origin. 2. Modifications You are allowed to modify the source code of the Software to enhance it. Modified versions must retain the original copyright notice and attribution. 3. "As Is" Warranty THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
load-testing, k6, performance, benchmark, go, python, scriptable
[ "Programming Language :: Python :: 3", "Programming Language :: Go", "Operating System :: OS Independent", "License :: Other/Proprietary License", "Topic :: Software Development :: Testing :: Traffic Generation" ]
[]
null
null
>=3.7
[]
[]
[]
[]
[]
[]
[]
[ "Homepage, https://github.com/richard24se/Sthenos", "Bug Tracker, https://github.com/richard24se/Sthenos/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:53:49.391310
sthenos-0.1.32.tar.gz
15,057,416
45/84/46342ac79b06df71e201abf8c53ba38410837f9cb3f4777f6dccccdcd5a3/sthenos-0.1.32.tar.gz
source
sdist
null
false
e152f839beca44eb4a9e369c2d02e7a3
07327bb1aa1556dbba359859fee4cba00db33368c1e657718e2a7a10e73bd5ef
458446342ac79b06df71e201abf8c53ba38410837f9cb3f4777f6dccccdcd5a3
null
[ "LICENSE" ]
203
2.4
optiprofiler
1.0.dev0
Benchmarking optimization solvers
OptiProfiler: a platform for benchmarking optimization solvers ============================================================== |docs_badge| |codecov_badge| .. image:: https://github.com/optiprofiler/optiprofiler/actions/workflows/matlab-unit_test.yml/badge.svg :target: https://github.com/optiprofiler/optiprofiler/actions/workflows/matlab-unit_test.yml .. image:: https://github.com/optiprofiler/optiprofiler/actions/workflows/matlab-test_multi-os.yml/badge.svg :target: https://github.com/optiprofiler/optiprofiler/actions/workflows/matlab-test_multi-os.yml .. image:: https://github.com/optiprofiler/optiprofiler/actions/workflows/matlab-random_test.yml/badge.svg :target: https://github.com/optiprofiler/optiprofiler/actions/workflows/matlab-random_test.yml .. image:: https://github.com/optiprofiler/optiprofiler/actions/workflows/matlab-stress_test.yml/badge.svg :target: https://github.com/optiprofiler/optiprofiler/actions/workflows/matlab-stress_test.yml .. |docs_badge| image:: https://img.shields.io/readthedocs/optiprofiler/latest?logo=readthedocs&style=for-the-badge :target: http://www.optprof.com .. |codecov_badge| image:: https://img.shields.io/codecov/c/github/optiprofiler/optiprofiler?style=for-the-badge&logo=codecov :target: https://app.codecov.io/github/optiprofiler/optiprofiler/tree/main MATLAB version is available. ---------------------------- To install the MATLAB version, please do the following: 1. Clone the repository using the following command: .. code-block:: bash git clone https://github.com/optiprofiler/optiprofiler.git 2. In MATLAB, navigate to the folder where the source code is located, and you will see a file named ``setup.m``. Run the following command in the MATLAB command window: .. code-block:: matlab setup Python version is under development. ------------------------------------
text/x-rst
null
Cunxin Huang <cun-xin.huang@connect.polyu.hk>, "Tom M. Ragonneau" <tom.ragonneau@polyu.edu.hk>, Zaikun Zhang <zaikun.zhang@polyu.edu.hk>
null
Cunxin Huang <cun-xin.huang@connect.polyu.hk>, "Tom M. Ragonneau" <tom.ragonneau@polyu.edu.hk>, Zaikun Zhang <zaikun.zhang@polyu.edu.hk>
BSD 3-Clause License Copyright (c) 2023-2025, Cunxin Huang, Tom M. Ragonneau, and Zaikun Zhang Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: 1. Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. 2. Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. 3. Neither the name of the copyright holder nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
null
[ "Development Status :: 1 - Planning", "Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: BSD License", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Topic :: Scientific/Engineering", "Topic :: Scientific/Engineering :: Mathematics", "Topic :: Software Development", "Topic :: Software Development :: Testing" ]
[]
null
null
>=3.8
[]
[]
[]
[ "matplotlib>=3.4.0", "numpy>=1.17.0", "scipy>=1.10.0", "pandas>=1.1.0", "h5py>=3.0.0", "pypdf>=3.8.0", "numpydoc==1.7.0; extra == \"doc\"", "Sphinx==7.3.6; extra == \"doc\"", "sphinx-copybutton==0.5.2; extra == \"doc\"", "sphinx-book-theme==1.1.2; extra == \"doc\"", "sphinx-design==0.5.0; extra == \"doc\"", "sphinxcontrib-bibtex==2.6.2; extra == \"doc\"", "sphinxcontrib-matlabdomain; extra == \"doc\"", "pycutest>=1.6.1; extra == \"extra\"", "pytest; extra == \"tests\"", "pytest-cov; extra == \"tests\"" ]
[]
[]
[]
[ "homepage, https://www.optprof.com", "documentation, http://www.optprof.com", "source, https://github.com/optiprofiler/optiprofiler", "download, https://pypi.org/project/optiprofiler/#files", "tracker, https://github.com/optiprofiler/optiprofiler/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:52:37.603710
optiprofiler-1.0.dev0.tar.gz
3,207,989
48/6c/a3d957d75a6b1f67b8d0db918f8a8c7ec86c8da0056fd11106527e863e5f/optiprofiler-1.0.dev0.tar.gz
source
sdist
null
false
21dc1deb5331dd58f7ece4661dbf0414
5972eb9efbfac61065e753f00df712d79a05b54435dbab9dc9edf8812018f79b
486ca3d957d75a6b1f67b8d0db918f8a8c7ec86c8da0056fd11106527e863e5f
null
[ "LICENSE" ]
193
2.4
veeam-ports-mcp
1.0.0
MCP server for querying Veeam product network port requirements
# Veeam Ports MCP Server An MCP (Model Context Protocol) server that gives Claude structured access to Veeam product network port requirements. Query ports, generate topology diagrams, and produce firewall rule import files — all from natural language. Backed by the [Magic Ports](https://magicports.veeambp.com) API, covering 25+ Veeam products. ## Installation ### Claude Desktop Add to your Claude Desktop config: - **macOS:** `~/Library/Application Support/Claude/claude_desktop_config.json` - **Windows:** `%APPDATA%\Claude\claude_desktop_config.json` ```json { "mcpServers": { "veeam-ports": { "command": "uvx", "args": ["veeam-ports-mcp"] } } } ``` Requires [uv](https://docs.astral.sh/uv/getting-started/installation/) — a single binary install. `uvx` downloads and runs the package automatically with no repo clone needed. ### VS Code / Claude Code ```bash claude mcp add veeam-ports -- uvx veeam-ports-mcp ``` ### Development Install ```bash git clone https://github.com/shapedthought/veeam-ports-mcp.git cd veeam-ports-mcp uv sync ``` Claude Desktop config for a local dev install: ```json { "mcpServers": { "veeam-ports": { "command": "uv", "args": [ "run", "--directory", "/path/to/veeam-ports-mcp", "veeam-ports-mcp" ] } } } ``` ## Available Tools | Tool | Description | |------|-------------| | `list_products` | List all Veeam products with port data | | `list_services` | List available service roles for a product — call this before `generate_topology` or `generate_app_import` | | `get_product_ports` | Get all port requirements for a product | | `get_product_subheadings` | Get section headings for a product — use to find valid `exclude_subsections` values | | `search_ports` | Free-text keyword search across all products | | `search_by_port_number` | Find which products and services use a specific port | | `get_source_details` | Source services with their section groupings | | `get_enriched_ports` | Port data with LLM-parsed service metadata | | `generate_topology` | Resolve firewall rules between named servers in your environment | | `generate_app_import` | Generate a JSON import file for the Magic Ports frontend app | ## Topology & Import File Workflow 1. Call `list_services` to see available service roles for the product 2. Ask the user which servers they have and what roles each one serves 3. Call `generate_topology` or `generate_app_import` with the server definitions 4. Optionally exclude subsections (e.g. `CDP Components`) or specific ports with `exclude_subsections` / `exclude_ports` ``` User: "Generate firewall rules for my VBR v13 environment. I have a VBR server, two Linux proxies, a repo, and ESXi hosts behind vCenter." Claude: [calls list_services → generate_app_import] "Import file saved: ~/Documents/veeam-ports-exports/magic-ports-vbr-v13-import.json" ``` Generated files are saved to `~/Documents/veeam-ports-exports/` by default. ## Example Prompts - "What ports does VBR v13 need?" - "Which products use port 902 and why?" - "Show me all ports the backup server uses to talk to ESXi hosts" - "Generate firewall rules for my VBR environment — I have a VBR server, a Linux proxy, a repo server, and ESXi hosts managed by vCenter" - "Create a Magic Ports import file for my VB365 deployment, excluding the proxy ports" - "What ports does the Veeam ONE server need open?" ## Configuration | Environment Variable | Default | Description | |---------------------|---------|-------------| | `VEEAM_PORTS_API_BASE` | `https://magicports.veeambp.com/ports_server` | API base URL | | `VEEAM_PORTS_OUTPUT_DIR` | `~/Documents/veeam-ports-exports` | Directory for generated import files | ## Debugging Use the MCP Inspector to test tools interactively: ```bash npx @modelcontextprotocol/inspector uvx veeam-ports-mcp ``` ## License MIT
text/markdown
null
null
null
null
MIT
null
[]
[]
null
null
>=3.10
[]
[]
[]
[ "httpx>=0.27.0", "mcp[cli]>=1.26.0" ]
[]
[]
[]
[ "Homepage, https://magicports.veeambp.com", "Repository, https://github.com/shapedthought/veeam-ports-mcp", "Issues, https://github.com/shapedthought/veeam-ports-mcp/issues" ]
uv/0.9.28 {"installer":{"name":"uv","version":"0.9.28","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"macOS","version":null,"id":null,"libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}
2026-02-20T21:52:24.793674
veeam_ports_mcp-1.0.0.tar.gz
66,004
de/57/792308f81af6284aa61755460f798361c8682a48eb6355ac87fc03ec5c47/veeam_ports_mcp-1.0.0.tar.gz
source
sdist
null
false
e7c5ee5c50cd05c145d29f7d13edc865
9535462537d90aedea47585e5f4049e47c4c2314e723fc109de8117282ca5b3c
de57792308f81af6284aa61755460f798361c8682a48eb6355ac87fc03ec5c47
null
[ "LICENSE" ]
228
2.4
remove-json-keys
1.5.0
Simply remove JSON keys via CLI command.
<a id="top"></a> # > remove-json-keys <a href="https://pypistats.org/packages/remove-json-keys"> <img height=31 src="https://img.shields.io/pypi/dm/remove-json-keys?logo=pypi&color=af68ff&logoColor=white&labelColor=464646&style=for-the-badge"></img></a> <a href="https://github.com/adamlui/python-utils/releases/tag/remove-json-keys-1.5.0"> <img height=31 src="https://img.shields.io/badge/Latest_Build-1.5.0-32fcee.svg?logo=icinga&logoColor=white&labelColor=464646&style=for-the-badge"></a> <a href="https://github.com/adamlui/python-utils/blob/main/remove-json-keys/docs/LICENSE.md"> <img height=31 src="https://img.shields.io/badge/License-MIT-f99b27.svg?logo=internetarchive&logoColor=white&labelColor=464646&style=for-the-badge"></a> <a href="https://www.codefactor.io/repository/github/adamlui/python-utils"> <img height=31 src="https://img.shields.io/codefactor/grade/github/adamlui/python-utils?label=Code+Quality&logo=codefactor&logoColor=white&labelColor=464646&color=a0fc55&style=for-the-badge"></a> <a href="https://sonarcloud.io/component_measures?metric=new_vulnerabilities&id=adamlui_python-utils"> <img height=31 src="https://img.shields.io/badge/dynamic/json?url=https%3A%2F%2Fsonarcloud.io%2Fapi%2Fmeasures%2Fcomponent%3Fcomponent%3Dadamlui_python-utils%26metricKeys%3Dvulnerabilities&query=%24.component.measures.0.value&style=for-the-badge&logo=sonarcloud&logoColor=white&labelColor=464646&label=Vulnerabilities&color=fafc74"></a> > ### _Simply remove JSON keys via CLI command._ ## Installation ```bash pip install remove-json-keys ``` ## Usage Run the CLI: ```bash remove-json-keys [options] # or remove-json ``` If no options are passed, the CLI will: 1. Prompt for keys to delete 2. Auto-discover closest child `json_dir` 3. Delete keys from found JSON files _Note: Key/values can span multiple lines and have any amount of whitespace between symbols._ ## Options Options can be set by using command-line arguments: | Option | Description | Example | ---------------------- | ------------------------------------------------------------------------------- | ----------------------------- | `-d`, `--json-dir` | Name of the folder containing JSON files (default: `_locales`) | `--json-dir=data` | `-k`, `--keys` | Comma-separated list of keys to remove | `--keys=app_DESC,err_NOT_FOUND` | `--config` | Use custom config file | `--config=path/to/file` | `init`, `-i`, `--init` | Create .remove-json.config.json5 in project root to store default settings | | `-n`, `--no-wizard` | Skip interactive prompts during start-up | | `-h`, `--help` | Show help screen | | `-v`, `--version` | Show version | | `--docs` | Open docs URL | ## Examples Remove `author` key from JSON files found in default `_locales` dir: ```bash remove-json-keys --keys=author # prompts for more keys to remove ``` Remove `info_SUCCESS` key from JSON files found in `messages` dir: ```bash remove-json-keys -n --keys=err_NOT_FOUND --json-dir=messages # no prompts ``` Remove `app_DESC` + `app_VER` keys from JSON files found in `data` dir: ```bash remove-json -n -k app_DESC,app_VER -d data # no prompts ``` ## Config file Run `remove-json init` to create `.remove-json.config.json5` in your project root to set default options. Example defaults: ```json5 { "json_dir": "_locales", // name of the folder containing JSON files "keys": "", // keys to remove (e.g. "app_NAME,author") "force": false, // force overwrite existing config file when using init "no_wizard": false // skip interactive prompts during start-up } ``` _Note: CLI arguments always override config file._ ## MIT License **Copyright © 2023–2026 [Adam Lui](https://github.com/adamlui).** Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. # <a href="#top">Back to top ↑</a>
text/markdown
null
Adam Lui <adam@kudoai.com>
null
null
null
cli, console, data, dev tool, json
[ "Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.15", "Topic :: File Formats :: JSON", "Topic :: Software Development :: Build Tools", "Topic :: Utilities" ]
[]
null
null
<4.0,>=3.6
[]
[]
[]
[ "colorama<1.0.0,>=0.4.6; platform_system == \"Windows\"", "json5<1.0.0,>=0.9.0", "nox>=2026.2.9; extra == \"dev\"", "tomli<3.0.0,>=2.0.0; extra == \"dev\"", "tomli-w<2.0.0,>=0.1.0; extra == \"dev\"" ]
[]
[]
[]
[ "Changelog, https://github.com/adamlui/python-utils/releases/tag/remove-json-keys-1.5.0", "Documentation, https://github.com/adamlui/python-utils/tree/main/remove-json-keys/docs", "Funding, https://github.com/sponsors/adamlui", "Homepage, https://github.com/adamlui/python-utils/tree/main/remove-json-keys/#readme", "Issues, https://github.com/adamlui/python-utils/issues", "PyPI Stats, https://pypistats.org/packages/remove-json-keys", "Releases, https://github.com/adamlui/python-utils/releases", "Repository, https://github.com/adamlui/python-utils" ]
twine/6.2.0 CPython/3.11.2
2026-02-20T21:52:17.851533
remove_json_keys-1.5.0.tar.gz
14,723
a7/f4/fe07613ea9a68cbe9d377713420ddb1bcda7be85eca1186ce7470c1b9fbb/remove_json_keys-1.5.0.tar.gz
source
sdist
null
false
d275a8d712b9d135218695595f9c29a8
bbe4f98b0d07e530d7979f1d4ad7829d302501a9ef13668053ee1f09d1ee1d2b
a7f4fe07613ea9a68cbe9d377713420ddb1bcda7be85eca1186ce7470c1b9fbb
MIT
[ "docs/LICENSE.md" ]
205
2.4
ty
0.0.18
An extremely fast Python type checker, written in Rust.
# ty [![ty](https://img.shields.io/endpoint?url=https://raw.githubusercontent.com/astral-sh/ty/main/assets/badge/v0.json)](https://github.com/astral-sh/ty) [![PyPI](https://img.shields.io/pypi/v/ty.svg)](https://pypi.python.org/pypi/ty) [![Discord](https://img.shields.io/badge/Discord-%235865F2.svg?logo=discord&logoColor=white)](https://discord.com/invite/astral-sh) An extremely fast Python type checker and language server, written in Rust. <br /> <p align="center"> <img alt="Shows a bar chart with benchmark results." width="500px" src="https://raw.githubusercontent.com/astral-sh/ty/main/docs/assets/ty-benchmark-cli-light.svg"> </p> <p align="center"> <i>Type checking the <a href="https://github.com/home-assistant/core">home-assistant</a> project without caching.</i> </p> <br /> ty is backed by [Astral](https://astral.sh), the creators of [uv](https://github.com/astral-sh/uv) and [Ruff](https://github.com/astral-sh/ruff). ty is currently in [beta](https://github.com/astral-sh/ty/blob/0.0.18/README.md#version-policy). ## Highlights - 10x - 100x faster than mypy and Pyright - Comprehensive [diagnostics](https://docs.astral.sh/ty/features/diagnostics/) with rich contextual information - Configurable [rule levels](https://docs.astral.sh/ty/rules/), [per-file overrides](https://docs.astral.sh/ty/reference/configuration/#overrides), [suppression comments](https://docs.astral.sh/ty/suppression/), and first-class project support - Designed for adoption, with support for [redeclarations](https://docs.astral.sh/ty/features/type-system/#redeclarations) and [partially typed code](https://docs.astral.sh/ty/features/type-system/#gradual-guarantee) - [Language server](https://docs.astral.sh/ty/features/language-server/) with code navigation, completions, code actions, auto-import, inlay hints, on-hover help, etc. - Fine-grained [incremental analysis](https://docs.astral.sh/ty/features/language-server/#fine-grained-incrementality) designed for fast updates when editing files in an IDE - Editor integrations for [VS Code](https://docs.astral.sh/ty/editors/#vs-code), [PyCharm](https://docs.astral.sh/ty/editors/#pycharm), [Neovim](https://docs.astral.sh/ty/editors/#neovim) and more - Advanced typing features like first-class [intersection types](https://docs.astral.sh/ty/features/type-system/#intersection-types), advanced [type narrowing](https://docs.astral.sh/ty/features/type-system/#top-and-bottom-materializations), and [sophisticated reachability analysis](https://docs.astral.sh/ty/features/type-system/#reachability-based-on-types) ## Getting started Run ty with [uvx](https://docs.astral.sh/uv/guides/tools/#running-tools) to get started quickly: ```shell uvx ty check ``` Or, check out the [ty playground](https://play.ty.dev) to try it out in your browser. To learn more about using ty, see the [documentation](https://docs.astral.sh/ty/). ## Installation To install ty, see the [installation](https://docs.astral.sh/ty/installation/) documentation. To add the ty language server to your editor, see the [editor integration](https://docs.astral.sh/ty/editors/) guide. ## Getting help If you have questions or want to report a bug, please open an [issue](https://github.com/astral-sh/ty/issues) in this repository. You may also join our [Discord server](https://discord.com/invite/astral-sh). ## Contributing Development of this project takes place in the [Ruff](https://github.com/astral-sh/ruff) repository at this time. Please [open pull requests](https://github.com/astral-sh/ruff/pulls) there for changes to anything in the `ruff` submodule (which includes all of the Rust source code). See the [contributing guide](https://github.com/astral-sh/ty/blob/0.0.18/CONTRIBUTING.md) for more details. ## Version policy ty uses `0.0.x` versioning. ty does not yet have a stable API; breaking changes, including changes to diagnostics, may occur between any two versions. See the [type system support](https://github.com/astral-sh/ty/issues/1889) tracking issue for a detailed overview of currently supported features. ## FAQ <!-- We intentionally use smaller headings for the FAQ items --> <!-- markdownlint-disable MD001 --> #### Why is ty doing \_\_\_\_\_? See our [typing FAQ](https://docs.astral.sh/ty/reference/typing-faq). #### How do you pronounce ty? It's pronounced as "tee - why" ([`/tiː waɪ/`](https://en.wikipedia.org/wiki/Help:IPA/English#Key)) #### How should I stylize ty? Just "ty", please. <!-- markdownlint-enable MD001 --> ## License ty is licensed under the MIT license ([LICENSE](https://github.com/astral-sh/ty/blob/0.0.18/LICENSE) or <https://opensource.org/licenses/MIT>). Unless you explicitly state otherwise, any contribution intentionally submitted for inclusion in ty by you, as defined in the MIT license, shall be licensed as above, without any additional terms or conditions. <div align="center"> <a target="_blank" href="https://astral.sh" style="background:none"> <img src="https://raw.githubusercontent.com/astral-sh/uv/main/assets/svg/Astral.svg" alt="Made by Astral"> </a> </div>
text/markdown; charset=UTF-8; variant=GFM
null
"Astral Software Inc." <hey@astral.sh>
null
null
null
ty, typing, analysis, check
[ "Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Developers", "Operating System :: OS Independent", "License :: OSI Approved :: MIT License", "Programming Language :: Python", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Rust", "Topic :: Software Development :: Quality Assurance", "Topic :: Software Development :: Testing", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules" ]
[]
https://github.com/astral-sh/ty/
null
>=3.8
[]
[]
[]
[]
[]
[]
[]
[ "Changelog, https://github.com/astral-sh/ty/blob/main/CHANGELOG.md", "Discord, https://discord.gg/astral-sh", "Releases, https://github.com/astral-sh/ty/releases", "Repository, https://github.com/astral-sh/ty" ]
uv/0.10.4 {"installer":{"name":"uv","version":"0.10.4","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
2026-02-20T21:51:42.039301
ty-0.0.18-py3-none-win32.whl
9,677,964
0d/da/f4ada0fd08a9e4138fe3fd2bcd3797753593f423f19b1634a814b9b2a401/ty-0.0.18-py3-none-win32.whl
py3
bdist_wheel
null
false
ae77bb10052cc94a3619050bb4cfede7
c5768607c94977dacddc2f459ace6a11a408a0f57888dd59abb62d28d4fee4f7
0ddaf4ada0fd08a9e4138fe3fd2bcd3797753593f423f19b1634a814b9b2a401
null
[ "LICENSE" ]
78,451
2.4
rwa-calc
0.1.25
Basel 3.1 Credit Risk RWA Calculator compliant with PRA PS9/24
*This package is still in development and is not production ready* # UK Credit Risk RWA Calculator [![Documentation](https://img.shields.io/badge/docs-GitHub%20Pages-blue)](https://OpenAfterHours.github.io/rwa_calculator/) A high-performance Risk-Weighted Assets (RWA) calculator for UK credit risk, supporting both current regulations and future Basel 3.1 implementation. Built with Python using Polars for vectorized performance. **Documentation:** [https://OpenAfterHours.github.io/rwa_calculator/](https://OpenAfterHours.github.io/rwa_calculator/) ## Installation ```bash # Install from PyPI pip install rwa-calc # Or with uv uv add rwa-calc # With UI support (web-based calculator interface) pip install rwa-calc[ui] ``` ### Optional Dependencies | Extra | Description | |-------|-------------| | `ui` | Interactive web UI via Marimo | | `dev` | Development tools (pytest, mypy, mkdocs) | | `all` | All optional dependencies | ## Quick Start **Option 1: Interactive UI** ```bash pip install rwa-calc[ui] rwa-calc-ui # Open http://localhost:8000 in your browser ``` **Option 2: Python API** ```python from datetime import date from rwa_calc.engine.pipeline import create_pipeline from rwa_calc.contracts.config import CalculationConfig config = CalculationConfig.crr(reporting_date=date(2026, 12, 31)) pipeline = create_pipeline() result = pipeline.run(config) print(f"Total RWA: {result.total_rwa:,.2f}") ``` ## Regulatory Scope This calculator supports two regulatory regimes: | Regime | Effective Period | UK Implementation | Status | |--------|------------------|-------------------|--------| | **CRR (Basel 3.0)** | Until 31 December 2026 | UK CRR (EU 575/2013 as onshored) | **Active Development** | | **Basel 3.1** | From 1 January 2027 | PRA PS9/24 | Planned | A configuration toggle allows switching between calculation modes for: - Current regulatory reporting under UK CRR - Impact analysis and parallel running ahead of Basel 3.1 go-live - Seamless transition when Basel 3.1 becomes effective ## Key Features - **Dual-Framework Support**: Single codebase for CRR and Basel 3.1 with UK-specific deviations - **High Performance**: Polars LazyFrames for vectorized calculations (50-100x improvement over row iteration) - **Complete Coverage**: Standardised (SA), IRB (F-IRB & A-IRB), and Slotting approaches - **Credit Risk Mitigation**: Collateral, guarantees, and provisions with RWA-optimized allocation - **Complex Hierarchies**: Multi-level counterparty and facility hierarchy support - **Audit Trail**: Full calculation transparency for regulatory review ### Supported Approaches | Approach | Description | |----------|-------------| | Standardised (SA) | Risk weights based on external ratings and exposure characteristics | | Foundation IRB (F-IRB) | Bank-estimated PD, supervisory LGD | | Advanced IRB (A-IRB) | Bank-estimated PD, LGD, and EAD | | Slotting | Category-based approach for specialised lending | ### Supported Exposure Classes Sovereign, Institution, Corporate, Corporate SME, Retail Mortgage, Retail QRRE, Retail Other, Specialised Lending, Equity ## Documentation Comprehensive documentation is available at **[OpenAfterHours.github.io/rwa_calculator](https://OpenAfterHours.github.io/rwa_calculator/)** | Section | Description | |---------|-------------| | [Getting Started](https://OpenAfterHours.github.io/rwa_calculator/getting-started/) | Installation and first calculation | | [User Guide](https://OpenAfterHours.github.io/rwa_calculator/user-guide/) | Regulatory frameworks, methodology, exposure classes | | [Architecture](https://OpenAfterHours.github.io/rwa_calculator/architecture/) | System design and pipeline | | [Data Model](https://OpenAfterHours.github.io/rwa_calculator/data-model/) | Input schemas and validation | | [API Reference](https://OpenAfterHours.github.io/rwa_calculator/api/) | Complete technical documentation | | [Development](https://OpenAfterHours.github.io/rwa_calculator/development/) | Testing, benchmarks, contributing | | [Plans](https://OpenAfterHours.github.io/rwa_calculator/plans/) | Development roadmap and status | ## Running Tests ```bash # Run all tests uv run pytest -v # Run with coverage uv run pytest --cov=src/rwa_calc # Run benchmarks uv run pytest tests/benchmarks/ -v ``` **Test Results:** 1,188 tests ## License [Apache-2.0 license](LICENSE) ## References ### Current Regulations (CRR / Basel 3.0) - [PRA Rulebook - CRR Firms](https://www.prarulebook.co.uk/pra-rules/crr-firms) - [UK CRR - Regulation (EU) No 575/2013 as onshored](https://www.legislation.gov.uk/eur/2013/575/contents) ### Basel 3.1 Implementation (January 2027) - [PRA PS9/24 - Implementation of the Basel 3.1 standards](https://www.bankofengland.co.uk/prudential-regulation/publication/2024/september/implementation-of-the-basel-3-1-standards-near-final-policy-statement-part-2) - [PRA CP16/22 - Implementation of Basel 3.1 Standards](https://www.bankofengland.co.uk/prudential-regulation/publication/2022/november/implementation-of-the-basel-3-1-standards) - [Basel Committee - CRE: Calculation of RWA for credit risk](https://www.bis.org/basel_framework/chapter/CRE/20.htm)
text/markdown
OpenAfterHours
null
OpenAfterHours
null
Apache-2.0
banking, basel, capital, credit-risk, polars, pra, regulatory, risk-management, rwa
[ "Development Status :: 2 - Pre-Alpha", "Intended Audience :: Developers", "Intended Audience :: Financial and Insurance Industry", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.13", "Topic :: Office/Business :: Financial", "Topic :: Scientific/Engineering :: Mathematics", "Typing :: Typed" ]
[]
null
null
>=3.13
[]
[]
[]
[ "duckdb>=0.9.0", "fastexcel>=0.19.0", "polars-normal-stats>=0.2.0", "polars>=1.0.0", "pyarrow>=14.0.0", "pydantic>=2.0.0", "pyyaml>=6.0.0", "marimo>=0.5.0; extra == \"all\"", "mkdocs-material>=9.5.0; extra == \"all\"", "mkdocs-mermaid2-plugin>=1.1.0; extra == \"all\"", "mkdocs>=1.5.0; extra == \"all\"", "mkdocstrings[python]>=0.24.0; extra == \"all\"", "mypy>=1.8.0; extra == \"all\"", "numpy>=1.26.0; extra == \"all\"", "pytest-benchmark>=4.0.0; extra == \"all\"", "pytest-cov>=4.0.0; extra == \"all\"", "pytest>=8.0.0; extra == \"all\"", "ruff>=0.3.0; extra == \"all\"", "uvicorn>=0.27.0; extra == \"all\"", "mkdocs-material>=9.5.0; extra == \"dev\"", "mkdocs-mermaid2-plugin>=1.1.0; extra == \"dev\"", "mkdocs>=1.5.0; extra == \"dev\"", "mkdocstrings[python]>=0.24.0; extra == \"dev\"", "mypy>=1.8.0; extra == \"dev\"", "numpy>=1.26.0; extra == \"dev\"", "pytest-benchmark>=4.0.0; extra == \"dev\"", "pytest-cov>=4.0.0; extra == \"dev\"", "pytest>=8.0.0; extra == \"dev\"", "ruff>=0.3.0; extra == \"dev\"", "marimo>=0.5.0; extra == \"ui\"", "uvicorn>=0.27.0; extra == \"ui\"" ]
[]
[]
[]
[ "Homepage, https://github.com/OpenAfterHours/rwa_calculator", "Documentation, https://OpenAfterHours.github.io/rwa_calculator/", "Repository, https://github.com/OpenAfterHours/rwa_calculator.git", "Issues, https://github.com/OpenAfterHours/rwa_calculator/issues", "Changelog, https://github.com/OpenAfterHours/rwa_calculator/blob/master/docs/appendix/changelog.md" ]
uv/0.9.21 {"installer":{"name":"uv","version":"0.9.21","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":null,"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}
2026-02-20T21:51:24.829522
rwa_calc-0.1.25.tar.gz
786,379
d9/5d/4f3152aff1c3f7b210d557213e2f38927ca05c3da0ae51c3855fb4b1552d/rwa_calc-0.1.25.tar.gz
source
sdist
null
false
95273c33e8ede54f09f8d847da662762
e0c61bc1be463b281dc54a28289bce060f46c6f7d2056e3b18519730de37fc4c
d95d4f3152aff1c3f7b210d557213e2f38927ca05c3da0ae51c3855fb4b1552d
null
[ "LICENSE" ]
210
2.4
deepagents-cli
0.0.25
Terminal interface for Deep Agents - interactive AI agent with file operations, shell access, and sub-agent capabilities.
# 🧠🤖 Deep Agents CLI [![PyPI - Version](https://img.shields.io/pypi/v/deepagents-cli?label=%20)](https://pypi.org/project/deepagents-cli/#history) [![PyPI - License](https://img.shields.io/pypi/l/deepagents-cli)](https://opensource.org/licenses/MIT) [![PyPI - Downloads](https://img.shields.io/pepy/dt/deepagents-cli)](https://pypistats.org/packages/deepagents-cli) [![Twitter](https://img.shields.io/twitter/url/https/twitter.com/langchain.svg?style=social&label=Follow%20%40LangChain)](https://x.com/langchain) <p align="center"> <img src="https://raw.githubusercontent.com/langchain-ai/deepagents/main/libs/cli/images/cli.png" alt="Deep Agents CLI" width="600"/> </p> ## Quick Install ```bash uv tool install deepagents-cli deepagents ``` ## 🤔 What is this? Using an LLM to call tools in a loop is the simplest form of an agent. This architecture, however, can yield agents that are "shallow" and fail to plan and act over longer, more complex tasks. Applications like "Deep Research", "Manus", and "Claude Code" have gotten around this limitation by implementing a combination of four things: a **planning tool**, **sub agents**, access to a **file system**, and a **detailed prompt**. `deepagents` is a Python package that implements these in a general purpose way so that you can easily create a Deep Agent for your application. For a full overview and quickstart of Deep Agents, the best resource is our [docs](https://docs.langchain.com/oss/python/deepagents/overview). **Acknowledgements: This project was primarily inspired by Claude Code, and initially was largely an attempt to see what made Claude Code general purpose, and make it even more so.** ## 📖 Resources - **[CLI Documentation](https://docs.langchain.com/oss/python/deepagents/cli/overview)** — Full documentation - **[CLI Source](https://github.com/langchain-ai/deepagents/tree/main/libs/cli)** — Full source code - **[Deep Agents SDK](https://github.com/langchain-ai/deepagents)** — The underlying agent harness - **[Chat LangChain](https://chat.langchain.com)** - Chat interactively with the docs ## 📕 Releases & Versioning See our [Releases](https://docs.langchain.com/oss/python/release-policy) and [Versioning](https://docs.langchain.com/oss/python/versioning) policies. ## 💁 Contributing As an open-source project in a rapidly developing field, we are extremely open to contributions, whether it be in the form of a new feature, improved infrastructure, or better documentation. For detailed information on how to contribute, see the [Contributing Guide](https://docs.langchain.com/oss/python/contributing/overview).
text/markdown
null
null
null
null
MIT
agents, ai, cli, deep-agent, langchain, langgraph, llm, terminal
[ "Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: Terminals" ]
[]
null
null
<4.0,>=3.11
[]
[]
[]
[ "aiosqlite<1.0.0,>=0.19.0", "daytona<1.0.0,>=0.113.0", "deepagents==0.4.3", "langchain-openai<2.0.0,>=1.1.8", "langchain<2.0.0,>=1.2.10", "langgraph-checkpoint-sqlite<4.0.0,>=3.0.0", "langsmith>=0.6.6", "markdownify<2.0.0,>=0.13.0", "modal<2.0.0,>=0.65.0", "pillow<13.0.0,>=10.0.0", "prompt-toolkit<4.0.0,>=3.0.52", "pyperclip<2.0.0,>=1.11.0", "python-dotenv<2.0.0,>=1.0.0", "pyyaml>=6.0.0", "requests<3.0.0,>=2.0.0", "rich<15.0.0,>=14.0.0", "runloop-api-client>=0.69.0", "tavily-python<1.0.0,>=0.7.21", "textual-autocomplete<5.0.0,>=3.0.0", "textual<9.0.0,>=8.0.0", "tomli-w<2.0.0,>=1.0.0", "langchain-anthropic<2.0.0,>=1.0.0; extra == \"all-providers\"", "langchain-aws<2.0.0,>=1.0.0; extra == \"all-providers\"", "langchain-cohere<1.0.0,>=0.5.0; extra == \"all-providers\"", "langchain-deepseek<2.0.0,>=1.0.0; extra == \"all-providers\"", "langchain-fireworks<2.0.0,>=1.0.0; extra == \"all-providers\"", "langchain-google-genai<5.0.0,>=4.0.0; extra == \"all-providers\"", "langchain-google-vertexai<4.0.0,>=3.0.0; extra == \"all-providers\"", "langchain-groq<2.0.0,>=1.0.0; extra == \"all-providers\"", "langchain-huggingface<2.0.0,>=1.0.0; extra == \"all-providers\"", "langchain-ibm<2.0.0,>=1.0.0; extra == \"all-providers\"", "langchain-mistralai<2.0.0,>=1.0.0; extra == \"all-providers\"", "langchain-nvidia-ai-endpoints<2.0.0,>=1.0.0; extra == \"all-providers\"", "langchain-ollama<2.0.0,>=1.0.0; extra == \"all-providers\"", "langchain-openai<2.0.0,>=1.1.8; extra == \"all-providers\"", "langchain-openrouter<2.0.0,>=0.0.1; extra == \"all-providers\"", "langchain-perplexity<2.0.0,>=1.0.0; extra == \"all-providers\"", "langchain-xai<2.0.0,>=1.0.0; extra == \"all-providers\"", "langchain-anthropic<2.0.0,>=1.0.0; extra == \"anthropic\"", "langchain-aws<2.0.0,>=1.0.0; extra == \"bedrock\"", "langchain-cohere<1.0.0,>=0.5.0; extra == \"cohere\"", "langchain-deepseek<2.0.0,>=1.0.0; extra == \"deepseek\"", "langchain-fireworks<2.0.0,>=1.0.0; extra == \"fireworks\"", "langchain-google-genai<5.0.0,>=4.0.0; extra == \"google-genai\"", "langchain-groq<2.0.0,>=1.0.0; extra == \"groq\"", "langchain-huggingface<2.0.0,>=1.0.0; extra == \"huggingface\"", "langchain-ibm<2.0.0,>=1.0.0; extra == \"ibm\"", "langchain-mistralai<2.0.0,>=1.0.0; extra == \"mistralai\"", "langchain-nvidia-ai-endpoints<2.0.0,>=1.0.0; extra == \"nvidia\"", "langchain-ollama<2.0.0,>=1.0.0; extra == \"ollama\"", "langchain-openai<2.0.0,>=1.1.8; extra == \"openai\"", "langchain-openrouter<2.0.0,>=0.0.1; extra == \"openrouter\"", "langchain-perplexity<2.0.0,>=1.0.0; extra == \"perplexity\"", "langchain-google-vertexai<4.0.0,>=3.0.0; extra == \"vertexai\"", "langchain-xai<2.0.0,>=1.0.0; extra == \"xai\"" ]
[]
[]
[]
[ "Homepage, https://docs.langchain.com/oss/python/deepagents/overview", "Documentation, https://reference.langchain.com/python/deepagents/", "Repository, https://github.com/langchain-ai/deepagents", "Issues, https://github.com/langchain-ai/deepagents/issues", "Changelog, https://github.com/langchain-ai/deepagents/blob/main/libs/cli/CHANGELOG.md", "Twitter, https://x.com/LangChain" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:50:53.604282
deepagents_cli-0.0.25.tar.gz
855,496
5e/b5/958e2bc48dcb865a5c53f483e46af5576450d1021613c17abb5d0f02ae03/deepagents_cli-0.0.25.tar.gz
source
sdist
null
false
f72271c9c8f7f6bb6e0c0d0b5578c879
480e3396d23c2c8bf925ce1060729f76f62b5aea820dee995095017d56ce9793
5eb5958e2bc48dcb865a5c53f483e46af5576450d1021613c17abb5d0f02ae03
null
[]
359
2.4
pyomo
6.10.0
The Pyomo optimization modeling framework
[![GitHub Actions Status](https://github.com/Pyomo/pyomo/actions/workflows/test_pr_and_main.yml/badge.svg?branch=main&event=push)](https://github.com/Pyomo/pyomo/actions/workflows/test_pr_and_main.yml?query=branch%3Amain+event%3Apush) [![Jenkins Status](https://github.com/Pyomo/jenkins-status/blob/main/pyomo_main.svg)](https://pyomo-jenkins.sandia.gov/) [![codecov](https://codecov.io/gh/Pyomo/pyomo/branch/main/graph/badge.svg)](https://codecov.io/gh/Pyomo/pyomo) [![Documentation Status](https://readthedocs.org/projects/pyomo/badge/?version=latest)](https://pyomo.readthedocs.org/en/latest/) [![Build services](https://github.com/Pyomo/jenkins-status/blob/main/pyomo_services.svg)](https://pyomo-jenkins.sandia.gov/) [![GitHub contributors](https://img.shields.io/github/contributors/pyomo/pyomo.svg)](https://github.com/pyomo/pyomo/graphs/contributors) [![Merged PRs](https://img.shields.io/github/issues-pr-closed-raw/pyomo/pyomo.svg?label=merged+PRs&color=44cc11)](https://github.com/pyomo/pyomo/pulls?q=is:pr+is:merged) [![a COIN-OR project](https://www.coin-or.org/GitHub/coin-or-badge.png)](https://www.coin-or.org) ## Pyomo Overview Pyomo is a Python-based open-source software package that supports a diverse set of optimization capabilities for formulating and analyzing optimization models. Pyomo can be used to define symbolic problems, create concrete problem instances, and solve these instances with standard solvers. Pyomo supports a wide range of problem types, including: - Linear programming - Quadratic programming - Nonlinear programming - Mixed-integer linear programming - Mixed-integer quadratic programming - Mixed-integer nonlinear programming - Mixed-integer stochastic programming - Generalized disjunctive programming - Differential algebraic equations - Mathematical programming with equilibrium constraints - Constraint programming Pyomo supports analysis and scripting within a full-featured programming language. Further, Pyomo has also proven an effective framework for developing high-level optimization and analysis tools. For example, the [`mpi-sppy`](https://github.com/Pyomo/mpi-sppy) package provides generic solvers for stochastic programming. `mpi-sppy` leverages the fact that Pyomo's modeling objects are embedded within a full-featured high-level programming language, which allows for transparent parallelization of subproblems using Python parallel communication libraries. * [Pyomo Home](https://www.pyomo.org) * [About Pyomo](https://www.pyomo.org/about) * [Download](https://www.pyomo.org/installation/) * [Documentation](https://www.pyomo.org/documentation/) * [Performance Plots](https://pyomo.github.io/performance/) Pyomo was formerly released as the Coopr software library. Pyomo is available under the BSD License - see the [`LICENSE.md`](https://github.com/Pyomo/pyomo/blob/main/LICENSE.md) file. Pyomo is currently tested with the following Python implementations: * CPython: 3.10, 3.11, 3.12, 3.13, 3.14 * PyPy: 3.11 _Testing and support policy_: At the time of the first Pyomo release after the end-of-life of a minor Python version, we will remove testing for that Python version. ### Installation #### PyPI &emsp; [![PyPI version](https://img.shields.io/pypi/v/pyomo.svg?maxAge=3600)](https://pypi.org/project/Pyomo/) [![PyPI downloads](https://img.shields.io/pypi/dm/pyomo.svg?maxAge=21600)](https://pypistats.org/packages/pyomo) pip install pyomo #### Anaconda &emsp; [![Anaconda version](https://img.shields.io/conda/vn/conda-forge/pyomo.svg)](https://anaconda.org/conda-forge/pyomo) [![Anaconda downloads](https://img.shields.io/conda/dn/conda-forge/pyomo.svg)](https://anaconda.org/conda-forge/pyomo) conda install -c conda-forge pyomo ### Tutorials and Examples * [Pyomo — Optimization Modeling in Python](https://link.springer.com/book/10.1007/978-3-030-68928-5) * [Pyomo Workshop Slides](https://github.com/Pyomo/pyomo-tutorials/blob/main/Pyomo-Workshop-December-2023.pdf) * [Prof. Jeffrey Kantor's Pyomo Cookbook](https://jckantor.github.io/ND-Pyomo-Cookbook/) * The [companion notebooks](https://mobook.github.io/MO-book/intro.html) for *Hands-On Mathematical Optimization with Python* * [Pyomo Gallery](https://github.com/Pyomo/PyomoGallery) ### Getting Help To get help from the Pyomo community ask a question on one of the following: * [Use the #pyomo tag on StackOverflow](https://stackoverflow.com/questions/ask?tags=pyomo) * [Pyomo Forum](https://groups.google.com/forum/?hl=en#!forum/pyomo-forum) ### Developers Pyomo development moved to this repository in June 2016 from Sandia National Laboratories. Developer discussions are hosted by [Google Groups](https://groups.google.com/forum/#!forum/pyomo-developers). The Pyomo Development team holds weekly coordination meetings on Tuesdays 12:30 - 14:00 (MT). Please contact wg-pyomo@sandia.gov to request call-in information. By contributing to this software project, you are agreeing to the following terms and conditions for your contributions: 1. You agree your contributions are submitted under the BSD license. 2. You represent you are authorized to make the contributions and grant the license. If your employer has rights to intellectual property that includes your contributions, you represent that you have received permission to make contributions and grant the required license on behalf of that employer. ### Related Packages See https://pyomo.readthedocs.io/en/latest/related_packages.html.
text/markdown
null
Pyomo Development Team <pyomo-developers@googlegroups.com>
null
null
null
optimization
[ "Development Status :: 5 - Production/Stable", "Intended Audience :: End Users/Desktop", "Intended Audience :: Science/Research", "Natural Language :: English", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: Unix", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Scientific/Engineering :: Mathematics", "Topic :: Software Development :: Libraries :: Python Modules" ]
[]
null
null
>=3.10
[]
[]
[]
[ "coverage; extra == \"tests\"", "parameterized; extra == \"tests\"", "pybind11; extra == \"tests\"", "pytest!=9.0.0; extra == \"tests\"", "pytest-parallel; extra == \"tests\"", "Sphinx!=8.2.0,!=9.0.*,!=9.1.0,>4; extra == \"docs\"", "sphinx-copybutton; extra == \"docs\"", "sphinx_rtd_theme>0.5; extra == \"docs\"", "sphinxcontrib-jsmath; extra == \"docs\"", "sphinxcontrib-napoleon; extra == \"docs\"", "numpy; extra == \"docs\"", "scipy; extra == \"docs\"", "dill; extra == \"optional\"", "ipython; extra == \"optional\"", "linear-tree; python_version < \"3.14\" and extra == \"optional\"", "scikit-learn<1.7.0; (implementation_name != \"pypy\" and python_version < \"3.14\") and extra == \"optional\"", "scikit-learn; (implementation_name != \"pypy\" and python_version >= \"3.14\") and extra == \"optional\"", "matplotlib!=3.6.1,>=3.6.0; extra == \"optional\"", "networkx; extra == \"optional\"", "numpy; extra == \"optional\"", "openpyxl; extra == \"optional\"", "packaging; extra == \"optional\"", "pint; implementation_name != \"pypy\" and extra == \"optional\"", "plotly; extra == \"optional\"", "python-louvain; extra == \"optional\"", "pyyaml; extra == \"optional\"", "qtconsole; extra == \"optional\"", "scipy; extra == \"optional\"", "sympy; extra == \"optional\"", "xlrd; extra == \"optional\"", "z3-solver; extra == \"optional\"", "pywin32; platform_system == \"Windows\" and extra == \"optional\"", "casadi; implementation_name != \"pypy\" and extra == \"optional\"", "numdifftools; implementation_name != \"pypy\" and extra == \"optional\"", "pandas; implementation_name != \"pypy\" and extra == \"optional\"", "seaborn; implementation_name != \"pypy\" and extra == \"optional\"" ]
[]
[]
[]
[ "Homepage, https://www.pyomo.org", "Documentation, https://pyomo.readthedocs.io/en/stable/", "Source, https://github.com/Pyomo/pyomo" ]
twine/6.2.0 CPython/3.12.8
2026-02-20T21:50:17.382727
pyomo-6.10.0-cp314-cp314t-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl
4,506,738
65/bd/cb4535b0bb63a7bcc968464ab739f54d2f4d15fe7924547336f5bd299513/pyomo-6.10.0-cp314-cp314t-manylinux_2_24_x86_64.manylinux_2_28_x86_64.whl
cp314
bdist_wheel
null
false
761a27bd9af88bde2f895c1374d48697
0590505b282468c9fc2555ced544213227a0a466bbc667826a36cc922afec6d7
65bdcb4535b0bb63a7bcc968464ab739f54d2f4d15fe7924547336f5bd299513
BSD-3-Clause
[ "LICENSE.md" ]
19,362
2.4
attest-ai
0.4.0
Test framework for AI agents
# attest-ai Test framework for AI agents. Deterministic assertions (schema validation, cost constraints, trace ordering, content matching) over agent execution traces. ## Install ```bash pip install attest-ai ``` With LLM provider support: ```bash pip install attest-ai[openai] # OpenAI pip install attest-ai[anthropic] # Anthropic pip install attest-ai[gemini] # Google Gemini pip install attest-ai[ollama] # Ollama (local) pip install attest-ai[all] # All providers ``` ## Quick start ```python import attest from attest import expect result = attest.AgentResult( trace=trace, # captured from your agent assertion_results=[], ) # Layer 1: Schema validation expect(result).output_matches_schema({"type": "object", "required": ["refund_id"]}) # Layer 2: Cost & performance constraints expect(result).cost_under(0.05) expect(result).latency_under(5000) # Layer 3: Trace structure expect(result).tools_called_in_order(["lookup_order", "process_refund"]) expect(result).no_tool_loops(max_iterations=3) # Layer 4: Content assertions expect(result).output_contains("refund") expect(result).output_not_contains("sorry") ``` ## Pytest integration Attest registers as a pytest plugin automatically: ```bash pytest tests/ --attest-engine=/path/to/attest-engine ``` ## Links - [Repository](https://github.com/attest-frameowrk/attest) - [Contributing](https://github.com/attest-frameowrk/attest/blob/main/CONTRIBUTING.md) - [License](https://github.com/attest-frameowrk/attest/blob/main/LICENSE) (Apache-2.0)
text/markdown
Attest Contributors
null
null
null
null
agents, ai, evaluation, llm, testing
[ "Development Status :: 3 - Alpha", "Framework :: Pytest", "Intended Audience :: Developers", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Testing" ]
[]
null
null
>=3.10
[]
[]
[]
[ "jsonschema>=4.20", "anthropic>=0.30; extra == \"all\"", "crewai>=0.60; extra == \"all\"", "google-adk>=1.0; extra == \"all\"", "google-genai>=1.0; extra == \"all\"", "langchain-core>=0.3; extra == \"all\"", "llama-index-core>=0.10.20; extra == \"all\"", "ollama>=0.4; extra == \"all\"", "openai>=1.30; extra == \"all\"", "opentelemetry-api>=1.20; extra == \"all\"", "opentelemetry-sdk>=1.20; extra == \"all\"", "anthropic>=0.30; extra == \"anthropic\"", "crewai>=0.60; extra == \"crewai\"", "build>=1.0; extra == \"dev\"", "mypy>=1.10; extra == \"dev\"", "pytest-asyncio>=0.23; extra == \"dev\"", "pytest>=8.0; extra == \"dev\"", "ruff>=0.4; extra == \"dev\"", "twine>=6.0; extra == \"dev\"", "google-genai>=1.0; extra == \"gemini\"", "google-adk>=1.0; extra == \"google-adk\"", "crewai>=0.60; extra == \"integration-test\"", "google-adk>=1.0; extra == \"integration-test\"", "google-genai>=1.0; extra == \"integration-test\"", "langchain-core>=0.3; extra == \"integration-test\"", "langchain-core>=0.3; extra == \"langchain\"", "llama-index-core>=0.10.20; extra == \"llamaindex\"", "ollama>=0.4; extra == \"ollama\"", "openai>=1.30; extra == \"openai\"", "opentelemetry-api>=1.20; extra == \"otel\"", "opentelemetry-sdk>=1.20; extra == \"otel\"" ]
[]
[]
[]
[ "Homepage, https://github.com/attest-frameowrk/attest", "Repository, https://github.com/attest-frameowrk/attest", "Issues, https://github.com/attest-frameowrk/attest/issues", "Documentation, https://github.com/attest-frameowrk/attest#readme" ]
twine/6.2.0 CPython/3.12.8
2026-02-20T21:49:49.111730
attest_ai-0.4.0.tar.gz
397,943
85/c5/dab3a2ed600835f0041f152981140fefb74f366a9646375e88105416bc09/attest_ai-0.4.0.tar.gz
source
sdist
null
false
905b0a05d63aa895008be00d408f9139
b9a9ea8f4f830da91c5e2fb60e3f930445cddf170992ca0b2fa244acf66112ae
85c5dab3a2ed600835f0041f152981140fefb74f366a9646375e88105416bc09
Apache-2.0
[]
211
2.4
cac-jira
0.6.5
A command-line interface for interacting with Jira
# Jira CLI A command-line interface for interacting with Jira. This project uses [UV](https://github.com/astral-sh/uv) for dependency management. ## Installation ```bash pip install cac-jira ``` ## Authentication On first-run, you'll be prompted for a Jira API token; generate one [here](https://id.atlassian.com/manage-profile/security/api-tokens). This will be stored in your system credential store (e.g. Keychain on Mac OS) in an items called `cac-jira`. ## Configuration On first-run, a configuration file will be generated at `~/.config/cac_jira/config.yaml`. In this file you'll need to replace the values of `server` and `username` with appropriate values. ```yaml server: https://your-jira-instance.atlassian.net project: YOUR_PROJECT_KEY # Optional default project username: your.email@example.com ``` ## Usage The Jira CLI follows a command-action pattern for all operations: ```bash jira <command> <action> [options] ``` ### Global Options - `--verbose`: Enable debug output - `--output [table|json]`: Control output format (default table) - `--help`: Show command help <!-- --suppress-output: Hide command output --> <!-- --version: Display version information --> ### Examples #### Issue Commands List issues in a project: ```bash jira issue list --project PROJ ``` List issues with additional filtering: ```bash jira issue list --project PROJ ``` Create a new issue: ```bash jira issue create --project PROJ --type Task --title "Fix login bug" --description "Users can't log in" ``` Create a new issue of a type that requires custom fields: ```bash # # This assumes the name of the custom fields is "Custom Field One" and "Custom Field Two"; # the field name will be swapped to lower-case, and spaces replaced with underscores # jira issue create --project PROJ --type Custom\ Issue\ Type --title "Issue Title" --description "Issue description" \ --field custom_field_one custom_field_value \ --field custom_field_two custom_field_value ``` Create and assign to yourself: ```bash jira issue create --project PROJ --type Bug --title "Server crash" --assign ``` Create and immediately start work: ```bash jira issue create --project PROJ --type Story --title "Add login feature" --begin ``` Add an issue to an epic: ```bash jira issue create --project PROJ --type Task --title "Subtask" --epic PROJ-100 ``` Label an issue: ```bash jira issue label --issue ISSUE_KEY --labels label1,label2 ``` Transition an issue: ```bash jira issue begin --issue ISSUE_KEY # Start work jira issue close --issue ISSUE_KEY # Mark as complete ``` #### Project Commands List all projects: ```bash jira project list ``` Show a project: ```bash jira project show --name PROJ-123 ``` #### Advanced Examples Update an issue's title or description: ```bash jira issue update --issue ISSUE_KEY --title "New issue title" --description "new issue description" ``` Add a comment to an issue: ```bash jira issue comment --issue ISSUE_KEY --comment "This is a comment." ``` List all issue IDs matching a label: ```bash jira issue list --output json | jq -r '.[] | select(.Labels | contains("production")) | .ID' ``` ## Development ### Setup Development Environment ```bash # Install dependencies including dev dependencies uv sync # Activate the venv source .venv/bin/activate # Run tests uv run pytest ``` Please note that tests are still WIP ### Project Structure - `cac_jira/commands/` - Command implementations - `issue/` - Issue-related commands - `project/` - Project-related commands - `cac_jira/cli/` - CLI entry point and argument parsing ### Adding New Commands 1. Create a new action module in the appropriate command directory. 2. Define a class that inherits from the command's base class. 3. Implement `define_arguments()` and `execute()` methods.
text/markdown
null
Ryan Punt <ryan@mirum.org>
null
null
null
jira, cli, atlassian, project-management, command-lint, python, cli-tool
[ "Programming Language :: Python :: 3", "Operating System :: OS Independent" ]
[]
null
null
<4.0,>=3.10
[]
[]
[]
[ "cac-core<1.0.0,>=0.6.0", "tabulate>=0.9.0", "jira<4.0.0,>=3.10.5", "pyyaml>=6.0.2", "keyring>=25.5.0", "argcomplete>=3.6.2", "mypy>=1.3.0; extra == \"dev\"", "types-pyyaml>=6.0.12; extra == \"dev\"", "types-tabulate>=0.9.0; extra == \"dev\"", "pytest>=7.3.1; extra == \"test\"", "black<25.0,>=23.3; extra == \"lint\"", "isort>=5.12.0; extra == \"lint\"", "pylint>=2.17.0; extra == \"lint\"", "sphinx>=7.0.0; extra == \"docs\"", "sphinx-rtd-theme>=1.2.0; extra == \"docs\"", "mypy>=1.3.0; extra == \"all\"", "types-pyyaml>=6.0.12; extra == \"all\"", "types-tabulate>=0.9.0; extra == \"all\"", "pytest>=7.3.1; extra == \"all\"", "black<25.0,>=23.3; extra == \"all\"", "isort>=5.12.0; extra == \"all\"", "pylint>=2.17.0; extra == \"all\"", "sphinx>=7.0.0; extra == \"all\"", "sphinx-rtd-theme>=1.2.0; extra == \"all\"" ]
[]
[]
[]
[ "homepage, https://mirum.org/cac-jira/", "repository, https://github.com/rpunt/cac-jira" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:49:42.886723
cac_jira-0.6.5.tar.gz
20,473
24/8f/3fbe3e42a7e4936bc4454b401cc276dff19b5c545093863b9a62b7eb36c4/cac_jira-0.6.5.tar.gz
source
sdist
null
false
9621439770955651143475e4e6e065ed
44ea3cc1d88142301d9cc2c5c12dd880d31bdb0020d6cb2007a95cf0c30368dc
248f3fbe3e42a7e4936bc4454b401cc276dff19b5c545093863b9a62b7eb36c4
MIT
[ "LICENSE" ]
200
2.3
szrpc
2026.2.3
Simple ZeroMQ RPC in Python
======================================= Swift RPC - Simple ZeroMQ RPC in Python ======================================= Overview ======== Swift RPC (szrpc) is a framework for creating remote python servers and and clients able to connect to them. It uses ZeroMQ for socket communications, and MessagePack for serialization. The key features which distinguish it from other existing solutions are: - Simple and clean API for creating clients, servers - Servers can support one or more workers running on the same host or many distinct hosts, with transparent load balancing - Supports multiple replies per request. Can be used to report progress for long running tasks or simply to send replies in chunks if the application needs it. - Reply objects can be transparently integrated into Gtk or Qt graphical frameworks through signals. Getting Started =============== Installing inside a virtual environment as follows :: $ python -m venv myproject $ source myproject/bin/activate (myproject) $ pip install szrpc If you plan to use the server dashboard, then install the `dash` extras by replacing the last command with: :: (myproject) $ pip install szrpc[dash] Write your first RPC Service ============================ The following example illustrates how simple it is to create one. .. code-block:: python from szrpc.server import Service class MyService(Service): def remote__hello_world(self, request, name=None): """ Single reply after a long duration """ request.reply(f'Please wait, {name} ...') time.sleep(10) return f'Hello, {name}. How is your world today?' def remote__date(self, request): """ Single reply after a short duration """ time.sleep(0.1) return f"Today's date is {datetime.now()}" def remote__progress(self, request): """ Multiple replies to show progress """ for i in range(10): time.sleep(0.1) request.reply(f'{i*10}% complete') return f"Progress done" The above example demonstrates the following key points applicable to Services: - Sevices must be sub-classes of **szrpc.server.Service**. - All methods prefixed with a `remote__` will be exposed remotely. - the very first argument to all remote methods is a request instance which contains all the information about the request. - The remaining arguments where present, must be keyword arguments. Positional arguments other than the initial `request` are not permitted. - Remote methods may block. - Multiple replies can be send back before the method completes. The return value will be the final reply sent to the client. Running a Server instance ------------------------- Once a service is defined, it can easily be used to start a server which can listen for incoming connections from multiple clients as follows: .. code-block:: python from szrpc.server import Server if __name__ == '__main__': service = MyService() server = Server(service=service, ports=(9990, 9991)) server.run() This says that our server will be available at the TCP address 'tcp://localhost:9990' for clients, and at the address 'tcp://localhost:9991' for workers. For simple cases, you don't need to worry about workers but by default, one worker created behind the scenes to provide the service, thus it is mandatory to specify both ports. Additionally, you can change your mind and run additional workers at any point in the future on any host after the server is started. To start the server with more than one worker on the local host, modify the `instances` keyword argument as follows: .. code-block:: python server = Server(service_factory=factory, ports=(9990, 9991), instances=1) It is possible to start the server with `instances = 0` however, it will obviously not be able to handle any requests until at least one worker is started. Server Dashboard ================ Swift RPC provides a web-based introspection capability that allows administrators to see active and historical calls, error messages, and basic statistics about the server. To enable introspection, make sure the [dash] extras are installed (`pip install szrpc[dash]`), then provide a `monitor_port` when creating the server: .. code-block:: python server = Server(service_factory=factory, ports=(9990, 9991), monitor_port=8080) server.run() Once the server is running, you can access the introspection dashboard by navigating to `http://localhost:8080` in your web browser. The dashboard provides: - **Uptime**: How long the server has been running. - **Total Requests**: Total number of requests processed. - **Errors**: Number of failed requests. - **Active Workers**: Number of currently connected workers. - **Active Calls**: List of requests currently being processed. - **Historical Calls**: List of recently completed requests with their call arguments, results or error messages. Starting External Workers ------------------------- Starting external workers is very similar to starting Servers. .. code-block:: python from szrpc import log from szrpc.server import Server, Service, WorkerManager from test_server import MyService if __name__ == '__main__': service = MyService() log.log_to_console() server = WorkerManager( service=service, backend="tcp://localhost:9991", instances=2 ) server.run() In the above example, we are staring two instances of workers on this host which are connected to the backend address of the main server. Creating Clients ---------------- Clients are just as easy, if not easier to create. Here is a test client for the above service. .. code-block:: python import time from szrpc import log from szrpc.client import Client # Define response handlers def on_done(res, data): print(f"Done: {res} {data!r}") def on_err(res, data): print(f"Failed: {res} : {data!r}") def on_update(res, data): print(f"Update: {res} {data!r}") if __name__ == '__main__': log.log_to_console() client = Client('tcp://localhost:9990') # wait for client to be ready before sending commands while not client.is_ready(): time.sleep(.001) res = client.hello_world(name='Joe') res.connect('done', on_done) res.connect('update', on_update) res.connect('failed', on_err) Here we have defined a few handler functions to get called once the replies are received. A few things are noteworthy in the above client code: - The client automatically figures out from the server, which methods to generate. For this reason, you will get "InvalidAttribute" errors if the initial handshake has not completed before method calls are made. For most production situations, this is not a problem but in the example above, we wait until the `client.is_ready()` returns `True` before proceeding. - The method names at the client end do not nave the `remote__` prefix. This means, overriding remote methods in the client will clobber the name. - Only key-worded arguments are allowed for remote methods. - Results are delivered asynchronously. To write synchronous code, you can call the `res.wait()` method on `Result` objects. There are three signal types corresponding to the three types of replies a server can send: 'done' the server has completed processing the request, no further replies should be expected for this request 'update' partial data is has been received for the request. More replies should be expected. 'failed' The request has failed. No more replies should be expected. Handler functions take two arguments, the first is always the `result` object, which is an instance of **szrpc.result.Result**, and the second is the decoded message from the server. Result Classes -------------- All results are instances of **szrpc.result.Result** or sub-classes thereof. The types of result objects produced can be changed to allow better integration with various frameworks. Presently, alternatives are available Gtk, Qt as well as a pure Python-based class. The pure Python result class is the default but it can easily be changed as follows. .. code-block:: python from szrpc.result.gresult import GResult import szrpc.client szrpc.client.use(GResult) my_client = szrpc.client.Client('tcp://localhost:9990') All subsequent result objects will be proper GObjects usable with the Gtk Main loop.
text/x-rst
Michel Fodje
michel.fodje@lightsource.ca
null
null
MIT
RPC, ZeroMQ, Networking, Development
[ "Intended Audience :: Developers", "Programming Language :: Python :: 3", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent" ]
[]
https://github.com/michel4j/swift-rpc
null
<4.0,>=3.9
[]
[]
[]
[ "pyzmq", "msgpack", "fastapi<0.130.0,>=0.127.0; extra == \"dash\"", "uvicorn<0.42.0,>=0.39.0; extra == \"dash\"" ]
[]
[]
[]
[ "Homepage, https://github.com/michel4j/swift-rpc", "Issues, https://github.com/michel4j/swift-rpc/issues" ]
poetry/2.1.1 CPython/3.13.11 Linux/6.18.6-200.fc43.x86_64
2026-02-20T21:49:14.012092
szrpc-2026.2.3.tar.gz
27,662
03/a9/185a3446e8c6df118db73f0a1be491351af97031d97a24086d38eb2a3ecd/szrpc-2026.2.3.tar.gz
source
sdist
null
false
dcbc767c141dc5559c8af55be56c0ab7
ad3dae292190e0b1c6b22620df150b8fbce19ebd9618a1522b66915a64dcf43c
03a9185a3446e8c6df118db73f0a1be491351af97031d97a24086d38eb2a3ecd
null
[]
202
2.1
ipystream
0.1.16
Easy interactive Jupyter dashboards, flowing top to bottom like a stream
# ipystream Easy interactive Jupyter dashboards, flowing top to bottom like a stream `python -m pytest` `poetry run black .` `poetry publish --build` To see poetry-repository-pypi token, in terminal run: `seahorse`
text/markdown
Charles Dabadie
dabadich@gmail.com
null
null
MIT
jupyter, dashboard, interactive, stream
[ "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13" ]
[]
https://github.com/jleblanc64/ipystream
null
>=3.10
[]
[]
[]
[ "ipydatagrid==1.1.16", "solara==1.45.0", "pydantic==2.11.7", "plotly==6.2.0", "PyJWT==2.10.1", "voila==0.5.10", "ipykernel==6.30.1" ]
[]
[]
[]
[ "Repository, https://github.com/jleblanc64/ipystream" ]
poetry/1.8.5 CPython/3.11.9 Linux/6.8.0-94-generic
2026-02-20T21:49:05.898824
ipystream-0.1.16.tar.gz
108,209
29/26/3434b9ad1691e322affa0d5d028321608744910c0b97e6d223f17b35e90b/ipystream-0.1.16.tar.gz
source
sdist
null
false
1cf04832d0c8205c09ca9a755739b12b
955e624fc5d36191538bf31c3fde742e304fc219ab0e13e1d83b0e3174007b7f
29263434b9ad1691e322affa0d5d028321608744910c0b97e6d223f17b35e90b
null
[]
207
2.4
smello
0.3.1
Capture outgoing HTTP requests and inspect them in a local web dashboard
# Smello Capture outgoing HTTP requests from your Python code and browse them in a local web dashboard — including gRPC calls made by Google Cloud libraries. Like [Mailpit](https://mailpit.axllent.org/), but for HTTP requests. ## Setup Install the client SDK and the server: ```bash pip install smello smello-server ``` Start the server: ```bash smello-server run ``` Add two lines to your code: ```python import smello smello.init() import requests resp = requests.get("https://api.stripe.com/v1/charges") # Browse captured requests at http://localhost:5110 ``` Smello monkey-patches `requests`, `httpx`, and `grpc` to capture all outgoing traffic. Browse results at `http://localhost:5110`. ### Google Cloud libraries Many Google Cloud Python libraries — BigQuery, Firestore, Pub/Sub, Analytics Data API (GA4), Vertex AI, Speech-to-Text, Vision, Translation, and others — use gRPC under the hood. Smello captures these calls automatically: ```python import smello smello.init() from google.cloud import bigquery client = bigquery.Client() rows = client.query("SELECT 1").result() # gRPC calls to bigquery.googleapis.com appear at http://localhost:5110 ``` Any library that calls `grpc.secure_channel()` or `grpc.insecure_channel()` is automatically captured. ## What Smello Captures - Method, URL, headers, and body - Response status code, headers, and body - Duration in milliseconds - Library used (requests, httpx, or grpc) Smello redacts sensitive headers (`Authorization`, `X-Api-Key`) by default. ## Configuration ```python smello.init( server_url="http://localhost:5110", # where to send captured data capture_hosts=["api.stripe.com"], # only capture these hosts capture_all=True, # capture everything (default) ignore_hosts=["localhost"], # skip these hosts redact_headers=["Authorization"], # replace values with [REDACTED] enabled=True, # kill switch ) ``` All parameters fall back to `SMELLO_*` environment variables when not passed explicitly: | Parameter | Env variable | Default | |-----------|-------------|---------| | `enabled` | `SMELLO_ENABLED` | `True` | | `server_url` | `SMELLO_URL` | `http://localhost:5110` | | `capture_all` | `SMELLO_CAPTURE_ALL` | `True` | | `capture_hosts` | `SMELLO_CAPTURE_HOSTS` | `[]` | | `ignore_hosts` | `SMELLO_IGNORE_HOSTS` | `[]` | | `redact_headers` | `SMELLO_REDACT_HEADERS` | `["Authorization", "X-Api-Key"]` | Boolean env vars accept `true`/`1`/`yes` and `false`/`0`/`no` (case-insensitive). List env vars are comma-separated. ## Supported Libraries - **requests** — patches `Session.send()` - **httpx** — patches `Client.send()` and `AsyncClient.send()` - **grpc** — patches `insecure_channel()` and `secure_channel()` to intercept unary-unary calls ## Requires - Python >= 3.10 - [smello-server](https://pypi.org/project/smello-server/) running locally ## Links - [Documentation & Source](https://github.com/smelloscope/smello) - [smello-server on PyPI](https://pypi.org/project/smello-server/)
text/markdown
null
Roman Imankulov <roman.imankulov@gmail.com>
null
null
null
null
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Topic :: Software Development :: Testing", "Topic :: System :: Networking :: Monitoring" ]
[]
null
null
>=3.10
[]
[]
[]
[]
[]
[]
[]
[ "Homepage, https://github.com/smelloscope/smello", "Repository, https://github.com/smelloscope/smello", "Issues, https://github.com/smelloscope/smello/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:49:04.836290
smello-0.3.1.tar.gz
15,331
d8/02/6bb3f8cdccf0032e60f4ef8bb320f516292ec0dad8d759afe1f033fc3ddd/smello-0.3.1.tar.gz
source
sdist
null
false
b34f5e09e6f9c037f8e5d155838e5cc0
8cc5fcf794e5d00875cc67fd13a08f179af9e8a0e9a1a068962b58d329051079
d8026bb3f8cdccf0032e60f4ef8bb320f516292ec0dad8d759afe1f033fc3ddd
MIT
[]
207
2.4
glyphh
0.4.2
Hyperdimensional Computing SDK and Runtime
# Glyphh Runtime Hyperdimensional computing runtime for deterministic, explainable AI. Glyphh encodes natural language into high-dimensional vector representations using Vector Symbolic Architecture (VSA). No LLM in the loop — just math. Same input, same output, every time. ## Features - **MCP Server** — Model Context Protocol interface for LLM sidecar integration - **GraphQL API** — Query knowledge graphs and fact trees with confidence scores - **CLI** — Manage models, deploy runtimes, and interact with the Glyphh Hub - **Deterministic** — Auditable, reproducible results grounded in cosine similarity ## Install Setup your python environment: ```bash python3 -m venv venv source venv/bin/activate ``` Glyphh ships as a single package with different install profiles: | Profile | Command | What you get | |---------|---------|-------------| | SDK | `pip install glyphh` | Encoder, similarity, CLI, model packaging. Lightweight — just numpy, pyyaml, click, httpx. | | Runtime | `pip install glyphh[runtime]` | Everything in SDK + FastAPI server, SQLAlchemy, pgvector, Alembic, Pydantic. For running the runtime locally. | | Dev | `pip install glyphh[dev]` | Everything in SDK + pytest, hypothesis, black, ruff, mypy. For contributing to Glyphh. | Most users want either SDK (build and package models) or Runtime (deploy and serve them). ## Quick Start The runtime requires PostgreSQL with pgvector. Pick whichever option fits your setup: ### Option 1 — Docker Compose (recommended) Requires [Docker Desktop](https://www.docker.com/products/docker-desktop/) (or Docker Engine + Compose plugin). The CLI can scaffold the Docker files for you: ```bash pip install glyphh[runtime] glyphh docker init docker pull ghcr.io/glyphh-ai/glyphh-runtime:latest docker compose up -d ``` `glyphh docker init` writes a `docker-compose.yml` and `init.sql` into your current directory. The compose file runs PostgreSQL with pgvector and the published runtime image — no build step needed. Verify it's running: ```bash curl http://localhost:8002/health ``` ### Option 2 — Docker (manual) Run the database and runtime as individual containers: ```bash docker run -d --name glyphh-db \ -e POSTGRES_USER=postgres \ -e POSTGRES_PASSWORD=postgres \ -e POSTGRES_DB=glyphh_runtime \ -p 5432:5432 \ pgvector/pgvector:pg16 docker pull ghcr.io/glyphh-ai/glyphh-runtime:latest docker run -p 8002:8002 \ -e DATABASE_URL=postgresql+asyncpg://postgres:postgres@host.docker.internal:5432/glyphh_runtime \ ghcr.io/glyphh-ai/glyphh-runtime:latest ``` Or with an existing database: ```bash docker run -p 8002:8002 \ -e DATABASE_URL=postgresql+asyncpg://user:pass@your-db-host:5432/glyphh \ ghcr.io/glyphh-ai/glyphh-runtime:latest ``` ### Option 3 — pip install (bring your own Postgres) If you already have PostgreSQL with pgvector running: ```bash pip install glyphh[runtime] export DATABASE_URL=postgresql+asyncpg://postgres:postgres@localhost:5432/glyphh_runtime glyphh serve ``` ### Query a deployed model ```bash glyphh query "What is the refund policy?" ``` ## How It Works 1. Your LLM sends a natural language query via MCP 2. Glyphh encodes it into a high-dimensional vector using stored procedures 3. The encoded query resolves against a knowledge graph via GraphQL 4. Fact trees with confidence scores are returned to ground the LLM's response ## License MIT
text/markdown
null
Glyphh AI <support@glyphh.ai>
null
null
MIT
null
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Scientific/Engineering :: Artificial Intelligence" ]
[]
null
null
>=3.9
[]
[]
[]
[ "numpy>=1.24.0", "pyyaml>=6.0", "click>=8.0.0", "httpx>=0.24.0", "fastapi>=0.100.0; extra == \"runtime\"", "uvicorn[standard]>=0.23.0; extra == \"runtime\"", "python-multipart>=0.0.6; extra == \"runtime\"", "sqlalchemy[asyncio]>=2.0.0; extra == \"runtime\"", "asyncpg>=0.28.0; extra == \"runtime\"", "pgvector>=0.2.0; extra == \"runtime\"", "alembic>=1.12.0; extra == \"runtime\"", "pydantic>=2.0.0; extra == \"runtime\"", "pydantic-settings>=2.0.0; extra == \"runtime\"", "email-validator>=2.0.0; extra == \"runtime\"", "pyjwt>=2.8.0; extra == \"runtime\"", "prometheus-client>=0.17.0; extra == \"runtime\"", "psutil>=5.9.0; extra == \"runtime\"", "pytest>=7.4.0; extra == \"dev\"", "pytest-asyncio>=0.21.0; extra == \"dev\"", "pytest-cov>=4.1.0; extra == \"dev\"", "hypothesis>=6.82.0; extra == \"dev\"", "aiosqlite>=0.19.0; extra == \"dev\"", "black>=23.7.0; extra == \"dev\"", "isort>=5.12.0; extra == \"dev\"", "mypy>=1.5.0; extra == \"dev\"", "ruff>=0.0.285; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://glyphh.ai", "Repository, https://github.com/glyphh-ai/glyphh-runtime", "Documentation, https://docs.glyphh.ai" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:48:51.533218
glyphh-0.4.2.tar.gz
227,055
e1/59/f7ec73e8b96e7cb0d1226f4daddee44fcfda1464bd4497567b3165cdb14e/glyphh-0.4.2.tar.gz
source
sdist
null
false
63c5a24bdcfefef9c8cd9a79bee383e8
b63d9d158ee6724d5596c89ae8341cabd57bad5b0c79ab1cda77bf995b738e87
e159f7ec73e8b96e7cb0d1226f4daddee44fcfda1464bd4497567b3165cdb14e
null
[ "LICENSE" ]
206
2.4
console-pong
1.0.0
A fully featured Pong game right in your terminal
Hi! This is CheezeDev. This is a very cool, playable Ping Pong game, which has trail effects, CPUs, etc! No need for dependencies. To insall, you can use pip install console-pong.
text/markdown
null
null
null
null
MIT
pong, game, terminal, console, cli
[ "Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: End Users/Desktop", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Topic :: Games/Entertainment :: Arcade" ]
[]
null
null
>=3.7
[]
[]
[]
[]
[]
[]
[]
[ "Homepage, https://github.com/CheezeDeveloper/console-pong" ]
twine/6.2.0 CPython/3.14.2
2026-02-20T21:48:31.141004
console_pong-1.0.0.tar.gz
9,267
b3/56/05a8ec57f3c8fb2c9813ff2486909234795be40c6221b8567e3c61ec76cd/console_pong-1.0.0.tar.gz
source
sdist
null
false
9526fa8b8d720607a63e01e03a084605
c8c572711c7bc01d63e8d1350405fbada59811f2831980de8103772d36c6900a
b35605a8ec57f3c8fb2c9813ff2486909234795be40c6221b8567e3c61ec76cd
null
[ "LICENSE" ]
227
2.4
certbot-dns-multi
4.32.1
Certbot DNS plugin supporting multiple providers, using github.com/go-acme/lego
# certbot-dns-multi [![certbot-dns-multi](https://snapcraft.io/certbot-dns-multi/badge.svg)](https://snapcraft.io/certbot-dns-multi) ![build status](https://github.com/alexzorin/certbot-dns-multi/actions/workflows/build-and-publish.yml/badge.svg) ![snap test status](https://github.com/alexzorin/certbot-dns-multi/actions/workflows/test-snap.yml/badge.svg) DNS plugin for [Certbot](https://certbot.eff.org/) which integrates with the 117+ DNS providers from the [`lego` ACME client](https://github.com/go-acme/lego/). At the last check, the supported providers are: > Akamai EdgeDNS, Alibaba Cloud DNS, all-inkl, Amazon Lightsail, Amazon Route 53, ArvanCloud, Aurora DNS, Autodns, Azure (deprecated), Azure DNS, Bindman, Bluecat, Brandit, Bunny, Checkdomain, Civo, Cloud.ru, CloudDNS, Cloudflare, ClouDNS, CloudXNS, ConoHa, Constellix, CPanel/WHM, Derak Cloud, deSEC.io, Designate DNSaaS for Openstack, Digital Ocean, DNS Made Easy, dnsHome.de, DNSimple, DNSPod (deprecated), Domain Offensive (do.de), Domeneshop, DreamHost, Duck DNS, Dyn, Dynu, EasyDNS, Efficient IP, Epik, Exoscale, External program, freemyip.com, G-Core, Gandi Live DNS (v5), Gandi, Glesys, Go Daddy, Google Cloud, Google Domains, Hetzner, Hosting.de, Hosttech, HTTP request, http.net, Hurricane Electric DNS, HyperOne, IBM Cloud (SoftLayer), IIJ DNS Platform Service, Infoblox, Infomaniak, Internet Initiative Japan, Internet.bs, INWX, Ionos, IPv64, iwantmyname, Joker, Joohoi's ACME-DNS, Liara, Linode (v4), Liquid Web, Loopia, LuaDNS, Mail-in-a-Box, Manual, Metaname, MyDNS.jp, MythicBeasts, Name.com, Namecheap, Namesilo, NearlyFreeSpeech.NET, Netcup, Netlify, Nicmanager, NIFCloud, Njalla, Nodion, NS1, Open Telekom Cloud, Oracle Cloud, OVH, plesk.com, Porkbun, PowerDNS, Rackspace, RcodeZero, reg.ru, RFC2136, RimuHosting, Sakura Cloud, Scaleway, Selectel, Servercow, Shellrent, Simply.com, Sonic, Stackpath, Tencent Cloud DNS, TransIP, UKFast SafeDNS, Ultradns, Variomedia, VegaDNS, Vercel, Versio.\[nl/eu/uk\], VinylDNS, VK Cloud, Vscale, Vultr, Webnames, Websupport, WEDOS, Yandex 360, Yandex Cloud, Yandex PDD, Zone.ee, Zonomi ## Installation ### via `snap` Using the `certbot` snap is the easiest way to use this plugin. See [here](https://certbot.eff.org/instructions?ws=other&os=snap) for instructions on installing Certbot via `snap`. ```bash sudo snap install certbot-dns-multi sudo snap set certbot trust-plugin-with-root=ok sudo snap connect certbot:plugin certbot-dns-multi ``` ### via `pip` Compiled wheels [are available](https://pypi.org/project/certbot-dns-multi/#files) for most `x86_64`/`amd64` Linux distributions. On other platforms, `pip` will try to compile the plugin, which requires [Go 1.19 or newer](https://go.dev/dl) to be installed on your server. | How did you install Certbot? | How to install the plugin | |-------------------------------------------------------------------------------------------------------|-------------------------------------------------------| | From `snap` | Don't use `pip`! Use the snap instructions above. | | Using the [official Certbot `pip` instructions](https://certbot.eff.org/instructions?ws=other&os=pip) | `sudo /opt/certbot/bin/pip install certbot-dns-multi` | | From `apt`, `yum`, `dnf` or any other distro package manager. (Requires Certbot 1.12.0 or newer.) | `pip install certbot-dns-multi` | ### via `docker` Docker images for `linux/amd64` and `linux/arm64` are available from [`ghcr.io/alexzorin/certbot-dns-multi`](https://ghcr.io/alexzorin/certbot-dns-multi). e.g. ```bash docker run --rm -it -v /etc/letsencrypt:/etc/letsencrypt \ ghcr.io/alexzorin/certbot-dns-multi certonly \ -a dns-multi --dns-multi-credentials /etc/letsencrypt/dns-multi.ini \ -d "*.example.com" -d "example.com" --dry-run ``` ## Usage `certbot-dns-multi` is controlled via a credentials file. 1. Head to https://go-acme.github.io/lego/dns/ and find your DNS provider in the list. In this example, we'll use `cloudflare`. 2. Create `/etc/letsencrypt/dns-multi.ini` and enter the name of your provider, all lowercase, as below: ```ini dns_multi_provider = cloudflare ``` 3. Following the instructions on https://go-acme.github.io/lego/dns/cloudflare/, we add the required configuration items: ```ini dns_multi_provider = cloudflare CLOUDFLARE_DNS_API_TOKEN="1234567890abcdefghijklmnopqrstuvwxyz" ``` 4. Save the file and secure it: ```bash chmod 0600 /etc/letsencrypt/dns-multi.ini ``` 5. Try issue a certificate now: ```bash certbot certonly -a dns-multi \ --dns-multi-credentials=/etc/letsencrypt/dns-multi.ini \ -d "*.example.com" \ --dry-run ``` 6. 🥳, or if not, ask on [the community forums](https://community.letsencrypt.org/) for help.
text/markdown
null
null
null
null
MIT
null
[]
[]
null
null
>=3.9
[]
[]
[]
[ "certbot>=1.12.0", "acme>=1.12.0", "josepy>=1.1.0", "black>=22.10.0; extra == \"dev\"", "flake8>=5.0.4; extra == \"dev\"" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:47:44.359283
certbot_dns_multi-4.32.1.tar.gz
73,233
20/0e/e5c6853af84c858075c0bbcdf26262c89d0d05eda1f4fe0783e3c28a9b06/certbot_dns_multi-4.32.1.tar.gz
source
sdist
null
false
10635e70b381e7e6fc96283427d0011f
32a0a13aa3f7c661ada6ce7d00fac18d0ba74f18c6b8138626c8ce7a6a0a20f3
200ee5c6853af84c858075c0bbcdf26262c89d0d05eda1f4fe0783e3c28a9b06
null
[ "LICENSE.txt" ]
22,694
2.3
bibtui
0.9.7
A quiet, powerful home for your references.
# bibtui > A quiet, powerful home for your references. [![PyPI](https://img.shields.io/pypi/v/bibtui)](https://pypi.org/project/bibtui/) [![Python 3.12+](https://img.shields.io/badge/python-3.12%2B-blue)](https://www.python.org/) [![License: MIT](https://img.shields.io/badge/license-MIT-green)](LICENSE) [![Publish](https://github.com/tgoelles/bib_tui/actions/workflows/publish.yml/badge.svg)](https://github.com/tgoelles/bib_tui/actions/workflows/publish.yml) [![CI](https://github.com/tgoelles/bib_tui/actions/workflows/ci.yml/badge.svg)](https://github.com/tgoelles/bib_tui/actions/workflows/ci.yml) ## Quick start ```bash # Run without installing uvx --prerelease=allow bibtui myrefs.bib # Or install permanently uv tool install --prerelease=allow bibtui ``` > **Why `--prerelease=allow`?** bibtui depends on `bibtexparser` v2, which is > still in beta on PyPI. This flag tells uv to use it. Once bibtexparser > publishes a stable v2 release this flag will no longer be needed. ----- ## Screenshots <!-- screenshots --> | Light theme | Dark — Catppuccin Mocha | | -------------------------------------- | ------------------------------------------- | | ![light theme](docs/pictures/theme1.png) | ![catppuccin mocha](docs/pictures/theme2.png) | | Nord — keywords modal | | | --------------------------------------------------- | - | | ![nord with keywords modal](docs/pictures/theme3.png) | | <!-- recording --> <!-- ![demo](docs/demo.gif) --> --- **bibtui** is a beautiful, keyboard-driven terminal app for researchers who live in the terminal. Browse and edit your `.bib` file, fetch open-access PDFs with a single keystroke, track what you've read, and never leave the command line — no database, no sync daemon, no account required. --- ## Why bibtui? | | bibtui | JabRef | Zotero | | -------------------------------- | ------ | ------- | ------ | | Runs in the terminal | ✅ | ❌ | ❌ | | No database / sync daemon | ✅ | ✅ | ❌ | | Git-friendly plain `.bib` | ✅ | ✅ | ❌ | | Works over SSH | ✅ | ❌ | ❌ | | Full Textual theming | ✅ | ❌ | ❌ | | Pure Python, installs in seconds | ✅ | ❌ | ❌ | --- ## Features - **Browse & search** — instant search across title, author, keywords, and cite key - **Import by DOI** — paste a DOI and metadata is fetched automatically - **Fetch PDFs automatically** — tries arXiv → Unpaywall (free, open-access) → direct URL - **Add existing PDFs** — pick a file from your Downloads folder with a live filter - **Edit entries** — field-by-field form *or* raw BibTeX editor (toggle with `v`) - **Read states & priorities** — track what you've read and what matters most - **Star ratings** — rate entries 1–5 - **Keywords editor** — manage tags inline - **JabRef-compatible** — file links use JabRef conventions; open the same `.bib` in both tools - **Git-friendly** — it's a plain text file (.bib); commit, diff, and collaborate normally - **Full Textual theme support** — including automatic detection of the [omarchy](https://omarchy.org) themes - **Works anywhere `uv` does** — SSH, HPC clusters, a colleague's laptop --- ## Installation ### Recommended — uv (fastest) ```bash uv tool install bibtui ``` ### pip ```bash pip install bibtui ``` ### Try without installing ```bash uvx bibtui references.bib ``` --- ## Usage ``` bibtui MyCollection.bib ``` On first launch bibtui shows a short onboarding wizard that pre-fills sensible defaults for your PDF directory, Downloads folder, and Unpaywall email (no registration required — the email is only used for rate-limiting). --- ## PDF workflow `f` tries three sources in order: 1. **arXiv** — for entries with a `10.48550/arXiv.*` DOI or an `arxiv.org` URL 2. **Unpaywall** — free open-access lookup by DOI (set your email in Settings; no account needed) 3. **Direct URL** — if the entry's `url` field points directly to a PDF PDFs are saved to your configured base directory and the entry's `file` field is updated automatically in JabRef format. --- ## Philosophy - Your `.bib` file is the source of truth - No hidden database - No setup, point and shoot possible - No lock-in - No accounts - keyboard and mouse support - nice looking - focused featurset. For cleanup use [bibtex-tidy](https://github.com/FlamingTempura/bibtex-tidy) or work directly on the bib file ## Development ```bash git clone https://github.com/tgoelles/bib_tui cd bib_tui uv sync uv run bibtui tests/bib_examples/MyCollection.bib ``` Run the tests: ```bash uv run pytest -m "not network" ``` Live-reload during development: ```bash uv run textual run --dev src/bibtui/main.py -- tests/bib_examples/MyCollection.bib ``` --- ## Related tools - [JabRef](https://www.jabref.org/) — GUI reference manager, same `.bib` format - [cobib](https://github.com/mrossinek/cobib) — another terminal BibTeX manager - [bibman](https://codeberg.org/KMIJPH/bibman) — minimal TUI reference manager --- ## FAQ **Does this modify my `.bib` formatting?** Yes. but we also write a backup file **Can I use it alongside JabRef?** Yes. File links follow JabRef conventions. ## License MIT © Thomas Gölles
text/markdown
Thomas Gölles
Thomas Gölles <thomas.goelles@gmail.com>
null
null
MIT License Copyright (c) 2026 Thomas Gölles Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
bibtex, bibliography, tui, terminal, latex
[ "Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "Intended Audience :: Information Technology", "Operating System :: OS Independent", "License :: OSI Approved :: MIT License", "Topic :: Text Processing :: Markup :: LaTeX", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14" ]
[]
null
null
>=3.12
[]
[]
[]
[ "bibtexparser>=2.0.0b9", "click>=8.3.1", "habanero>=2.3.0", "textual>=8.0.0" ]
[]
[]
[]
[ "Homepage, https://github.com/tgoelles/bib_tui", "Repository, https://github.com/tgoelles/bib_tui", "Bug Tracker, https://github.com/tgoelles/bib_tui/issues" ]
uv/0.10.4 {"installer":{"name":"uv","version":"0.10.4","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
2026-02-20T21:47:35.801804
bibtui-0.9.7.tar.gz
29,724
d8/9c/8c455e6be68dc790bed2bfbed600ad1604fb4972eb4fa00043f0fedf0d71/bibtui-0.9.7.tar.gz
source
sdist
null
false
95533249cc5e7de69f1b83631cf6f255
0f781bcba76850edad7bd492b2affba2994ce99964023341afc52a872a1b2461
d89c8c455e6be68dc790bed2bfbed600ad1604fb4972eb4fa00043f0fedf0d71
null
[]
210
2.4
translate-messages
1.5.0
Translate en/messages.json (in chrome.i18n format) to 100+ locales automatically.
<a id="top"></a> # > translate-messages <a href="https://pypistats.org/packages/translate-messages"> <img height=31 src="https://img.shields.io/pypi/dm/translate-messages?logo=pypi&color=af68ff&logoColor=white&labelColor=464646&style=for-the-badge"></img></a> <a href="https://github.com/adamlui/python-utils/releases/tag/translate-messages-1.5.0"> <img height=31 src="https://img.shields.io/badge/Latest_Build-1.5.0-32fcee.svg?logo=icinga&logoColor=white&labelColor=464646&style=for-the-badge"></a> <a href="https://github.com/adamlui/python-utils/blob/main/translate-messages/docs/LICENSE.md"> <img height=31 src="https://img.shields.io/badge/License-MIT-f99b27.svg?logo=internetarchive&logoColor=white&labelColor=464646&style=for-the-badge"></a> <a href="https://www.codefactor.io/repository/github/adamlui/python-utils"> <img height=31 src="https://img.shields.io/codefactor/grade/github/adamlui/python-utils?label=Code+Quality&logo=codefactor&logoColor=white&labelColor=464646&color=a0fc55&style=for-the-badge"></a> <a href="https://sonarcloud.io/component_measures?metric=new_vulnerabilities&id=adamlui_python-utils"> <img height=31 src="https://img.shields.io/badge/dynamic/json?url=https%3A%2F%2Fsonarcloud.io%2Fapi%2Fmeasures%2Fcomponent%3Fcomponent%3Dadamlui_python-utils%26metricKeys%3Dvulnerabilities&query=%24.component.measures.0.value&style=for-the-badge&logo=sonarcloud&logoColor=white&labelColor=464646&label=Vulnerabilities&color=fafc74"></a> > ### _Translate `en/messages.json` (in chrome.i18n format) to 100+ locales automatically._ ## Installation ```bash pip install translate-messages ``` ## Usage Run the CLI: ```bash translate-messages [options] # or translate-msgs ``` If no options are passed, the CLI will: 1. Prompt for message keys to ignore 2. Auto-discover closest child `_locales` dir 3. Translate found `en/messages.json` to target languages _Note: Any messages.json in the [`chrome.i18n`](https://developer.chrome.com/docs/extensions/how-to/ui/localization-message-formats) format can be used as a source file._ ## Options Options can be set by using command-line arguments: | Option | Description | Example | ---------------------- | --------------------------------------------------------------------------------------------------------- | ------------------------------ | `-d`, `--locales-dir` | Name of the folder containing locale files (default: `_locales`) | `--locales-dir=_messages` | `-t`, `--target-langs` | Comma-separated list of languages to translate to (default: all 100+ [`stable_locales`][stable-locales]) | `--target-langs=es,fr` | `-k`, `--keys` | Comma-separated list of keys to translate (default: all found src keys missing in target files) | `--keys=app_DESC,err_NOT_FOUND` | `--exclude-langs` | Comma-separated list of languages to exclude | `--exclude-langs=es,zh` | `--exclude-keys` | Comma-separated list of keys to ignore | `--exclude-keys=app_NAME,author` | `--only-stable` | Only use stable locales (skip auto-discovery) | | `--config` | Use custom config file | `--config=path/to/file` | `init`, `-i`, `--init` | Create `.translate-msgs.config.json5` in project root to store default options | | `-f`, `--force` | Force overwrite of existing config file when using `init` | | `-n`, `--no-wizard` | Skip interactive prompts during start-up | | `-h`, `--help` | Show help screen | | `-v`, `--version` | Show version | | `--docs` | Open docs URL | ## Examples Translate all keys except `app_NAME` from `_locales/en/messages.json` to all [`stable_locales`][stable-locales]: ```bash translate-messages --ignore-keys=app_NAME # prompts for more keys to ignore ``` Translate `app_DESC` key from `messges/en/messages.json` to French: ```bash translate-messages -n --keys=app_DESC --locales-dir=messages --target-langs=fr # no prompts ``` Translate `app_DESC` + `err_NOT_FOUND` keys from `_msgs/en/messages.json` to Spanish and Hindi: ```bash translate-msgs -n -k app_DESC,err_NOT_FOUND -d _msgs -t es,hi # no prompts ``` ## Config file Run `translate-msgs init` to create `.translate-msgs.config.json5` in your project root to set default options. Example defaults: ```json5 { "locales_dir": "_locales", // name of the folder containing locale files "target_langs": "", // languages to translate to (e.g. "en,es,fr") (default: all 100+ supported locales) "keys": "", // keys to translate (e.g. "app_DESC,err_NOT_FOUND") "exclude_langs": "", // languages to exclude (e.g. "en,es") "exclude_keys": "", // keys to ignore (e.g. "app_NAME,author") "force": false, // force overwrite existing config file when using init "no_wizard": false // skip interactive prompts during start-up } ``` _Note: CLI arguments always override config file._ ## MIT License **Copyright © 2023–2026 [Adam Lui](https://github.com/adamlui).** Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. # <a href="#top">Back to top ↑</a> [stable-locales]: https://github.com/adamlui/python-utils/blob/translate-messages-1.5.0/translate-messages/src/translate_messages/assets/data/package_data.json#L23-L28
text/markdown
null
Adam Lui <adam@kudoai.com>
null
null
null
chrome, cli, console, data, dev tool, i18n, internationalization, json, localization, messages, mymemory, translate, translation, translator
[ "Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: Science/Research", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.15", "Topic :: File Formats :: JSON", "Topic :: Software Development :: Build Tools", "Topic :: Software Development :: Internationalization", "Topic :: Software Development :: Localization", "Topic :: Utilities" ]
[]
null
null
<4.0,>=3.6
[]
[]
[]
[ "colorama<1.0.0,>=0.4.6; platform_system == \"Windows\"", "json5<1.0.0,>=0.9.0", "translate<4.0.0,>=3.8.0", "nox>=2026.2.9; extra == \"dev\"", "tomli<3.0.0,>=2.0.0; extra == \"dev\"", "tomli-w<2.0.0,>=0.1.0; extra == \"dev\"" ]
[]
[]
[]
[ "Changelog, https://github.com/adamlui/python-utils/releases/tag/translate-messages-1.5.0", "Documentation, https://github.com/adamlui/python-utils/tree/main/translate-messages/docs", "Funding, https://github.com/sponsors/adamlui", "Homepage, https://github.com/adamlui/python-utils/tree/main/translate-messages/#readme", "Issues, https://github.com/adamlui/python-utils/issues", "PyPI Stats, https://pypistats.org/packages/translate-messages", "Releases, https://github.com/adamlui/python-utils/releases", "Repository, https://github.com/adamlui/python-utils" ]
twine/6.2.0 CPython/3.11.2
2026-02-20T21:47:03.945644
translate_messages-1.5.0.tar.gz
17,847
65/f4/04bd0e635f272e2e33db7ca39d49c2f7befc37bddb2108fe40ec4e49aeb5/translate_messages-1.5.0.tar.gz
source
sdist
null
false
d2a7694960b06b7e38c8a064639819c5
700ab8701042da18f691ddfea542fb3c74adc0e9eb7101508db014d4a829f35b
65f404bd0e635f272e2e33db7ca39d49c2f7befc37bddb2108fe40ec4e49aeb5
MIT
[ "docs/LICENSE.md" ]
208
2.4
firthmodels
0.7.2
Firth-penalized models in Python
# firthmodels [![CI](https://github.com/jzluo/firthmodels/actions/workflows/ci.yml/badge.svg)](https://github.com/jzluo/firthmodels/actions/workflows/ci.yml) [![PyPI](https://img.shields.io/pypi/v/firthmodels)](https://pypi.org/project/firthmodels/) ![Pepy Total Downloads](https://img.shields.io/pepy/dt/firthmodels) [![Conda Version](https://img.shields.io/conda/vn/conda-forge/firthmodels)](https://anaconda.org/channels/conda-forge/packages/firthmodels/overview) ![Python Version from PEP 621 TOML](https://img.shields.io/python/required-version-toml?tomlFilePath=https%3A%2F%2Fraw.githubusercontent.com%2Fjzluo%2Ffirthmodels%2Frefs%2Fheads%2Fmain%2Fpyproject.toml) ![GitHub License](https://img.shields.io/github/license/jzluo/firthmodels) [![DOI](https://zenodo.org/badge/DOI/10.5281/zenodo.17863280.svg)](https://doi.org/10.5281/zenodo.17863280) Firth-penalized models in Python: - `FirthLogisticRegression`: scikit-learn–compatible Firth logistic regression - `FirthCoxPH`: scikit-survival-style Firth Cox proportional hazards Firth penalization reduces small-sample bias and produces finite estimates even when standard MLE fails due to (quasi-)complete separation or monotone likelihood. See [benchmarking results here](https://github.com/jzluo/firthmodels/blob/main/benchmarks/README.md) comparing firthmodels, [logistf](https://cran.r-project.org/web/packages/logistf/index.html), [brglm2](https://cran.r-project.org/web/packages/brglm2/index.html), and [coxphf](https://cran.r-project.org/web/packages/coxphf/index.html). ## Why Firth penalization? Standard maximum-likelihood logistic regression fails when your data has complete or quasi-complete separation: when a predictor (or combination of predictors) perfectly separates the outcome classes. In these cases, MLE produces infinite coefficient estimates. In Cox proportional hazards, an analogous failure mode is monotone likelihood, where the partial likelihood becomes unbounded (often due to small samples, rare events, or near-perfect risk separation). These problems are common in: - Case-control studies with rare exposures - Small clinical trials - Genome-wide or Phenome-wide association studies (GWAS/PheWAS) - Any dataset where events are rare relative to predictors Firth's method adds a penalty term that: - Produces **finite, well-defined estimates** even with separated data - **Reduces small-sample bias** in coefficient estimates Kosmidis and Firth (2021) formally proved that bias reduction for logistic regression models guarantees finite estimates as long as the model matrix has full rank. ### Detecting separation You can use `detect_separation` to check if your data has separation before fitting. This implements the linear programming method from Konis (2007), as used in the R detectseparation package by Kosmidis et al (2022). The following example is based on the endometrial dataset used in Heinze and Schemper (2002), where the `NV` feature causes quasi-complete separation. ```python from firthmodels import detect_separation result = detect_separation(X, y) result.separation # True result.is_finite # array([False, True, True, True]) result.directions # array([1, 0, 0, 0]) # where 1 = +Inf, -1 = -Inf, 0 = finite print(result.summary()) # Separation: True # NV +Inf # PI finite # EH finite # intercept finite ``` ## Installation ### Pip ```bash pip install firthmodels ``` Requires Python 3.10+ and depends on NumPy, SciPy, and scikit-learn. Optional dependencies: - Numba acceleration: `pip install firthmodels[numba]` - The first run with the Numba backend after installing or updating firthmodels may take 10-30 seconds due to JIT compilation. Subsequent runs are fast thanks to caching. - Formula interface for the statsmodels adapter: `pip install firthmodels[formula]` (or simply install [formulaic](https://matthewwardrop.github.io/formulaic/latest/)). **Note:** Performance is significantly improved when NumPy/SciPy are built against a well-optimized BLAS/LAPACK library. You can check which library yours is using with `np.show_config()`. As a rule of thumb, MKL offers the best performance for Intel CPUs, while OpenBLAS is also a good choice for Intel and generally the best option for AMD. On macOS, ensure NumPy/SciPy are linked to Apple Accelerate. The most straightforward way to control the BLAS/LAPACK library is to install `firthmodels` in a conda environment: ### conda ```bash conda install -c conda-forge firthmodels # usually defaults to OpenBLAS conda install -c conda-forge firthmodels "libblas=*=*_newaccelerate" # Apple Accelerate conda install -c conda-forge firthmodels "libblas=*=*mkl" # Intel MKL conda install -c conda-forge firthmodels "libblas=*=*openblas" # OpenBLAS ``` Add numba to the conda install command to enable Numba acceleration. ## Quick start ### Firth logistic regression ```python import numpy as np from firthmodels import FirthLogisticRegression # Separated data: x=1 perfectly predicts y=1 X = np.array([[0], [0], [0], [1], [1], [1]]) y = np.array([0, 0, 0, 1, 1, 1]) # Standard logistic regression would fail here model = FirthLogisticRegression().fit(X, y) print(model.coef_) # array([3.89181893]) print(model.intercept_) # -2.725... print(model.pvalues_) # Wald p-values print(model.bse_) # Standard errors ``` ### Firth Cox proportional hazards ```python import numpy as np from firthmodels import FirthCoxPH X = np.array([[1.0], [0.0]]) event = np.array([True, False]) time = np.array([1.0, 2.0]) model = FirthCoxPH().fit(X, (event, time)) print(model.coef_) # log hazard ratios print(model.pvalues_) # Wald p-values # Survival curves evaluated at the training event times S = model.predict_survival_function(X) # shape: (n_samples, n_event_times) ``` `FirthCoxPH` also accepts `y` as a structured array with boolean `event` and float `time` fields (scikit-survival style). Both estimators take a `backend` parameter that can be one of `'auto'` (default), `'numba'`, or `'numpy'`. If `'auto'`, firthmodels auto-detects Numba availability and uses it if installed, otherwise numpy/scipy. ## Estimators ### scikit-learn compatible API `FirthLogisticRegression` follows the scikit-learn estimator API (`fit`, `predict`, `predict_proba`, `get_params`, `set_params`, etc.), and can be used with pipelines, cross-validation, and other sklearn tools: ```python from sklearn.model_selection import cross_val_score from sklearn.pipeline import make_pipeline from sklearn.preprocessing import StandardScaler pipe = make_pipeline(StandardScaler(), FirthLogisticRegression()) scores = cross_val_score(pipe, X, y, cv=5) ``` `FirthCoxPH` also follows the sklearn estimator API (`fit`, `predict`, `score`, etc.). It also has a scikit-survival-like interface: - methods: `fit(X, y)`, `predict(X)` (linear predictor), `score(X, y)` (C-index), `predict_survival_function(X, return_array=True)`, `predict_cumulative_hazard_function(X, return_array=True)`. - attributes: `unique_times_`, `cum_baseline_hazard_`, `baseline_survival_` (Breslow-style baseline). ## Inference ### Likelihood ratio tests (LRT) Both estimators support penalized likelihood ratio tests for individual coefficients. These are often more reliable than Wald p-values in small samples. `lrt()` populates `lrt_pvalues_` and `lrt_bse_` (a back-corrected standard error such that `(beta / lrt_bse_)**2` matches the 1-df chi-squared statistic), which can be useful for meta-analysis weighting: ```python model.fit(X, y).lrt() # Compute LRT for all coefficients model.lrt_pvalues_ # LRT p-values model.lrt_bse_ # Back-corrected standard errors (separate from Wald bse_) ``` Each feature requires a separate constrained model fit, so you can test selectively to avoid unnecessary computation. By default, LRT uses a warm start based on the full-model covariance to reduce Newton-Raphson iterations; pass `warm_start=False` to disable it. ```python model.lrt(0) # Single feature by index model.lrt([0, 2]) # Multiple features model.lrt(['snp', 'age']) # By name (if fitted with DataFrame) model.lrt(['snp', 2]) # Mixed model.lrt(warm_start=False) # Disable warm start ``` ### Confidence intervals ```python model.conf_int() # 95% Wald CIs model.conf_int(alpha=0.1) # 90% CIs model.conf_int(method='pl') # Profile likelihood CIs (more accurate) model.conf_int(method='pl', features=['snp', 'age']) # can selectively compute as with LRT ``` ### Sample weights and offsets ```python # currently for FirthLogisticRegression only model.fit(X, y, sample_weight=weights) model.fit(X, y, offset=offset) ``` ## Statsmodels adapter (`FirthLogit`) The statsmodels adapter wraps `FirthLogisticRegression` behind a statsmodels-like API and returns a results object with common statsmodels attributes/methods (`params`, `bse`, `pvalues`, `summary()`, `cov_params()`, etc.). Notes: - Unlike sklearn, `FirthLogit` does not add an intercept automatically; use `sm.add_constant(X)`. - `fit(pl=True)` (default) computes likelihood ratio p-values and uses profile likelihood confidence intervals by default, matching R logistf convention. Standard errors (`bse`) remain Wald standard errors. ```python import numpy as np import statsmodels.api as sm from firthmodels.adapters.statsmodels import FirthLogit X = np.array([[1], [2], [3], [4], [5], [6]]) y = np.array([0, 0, 0, 1, 1, 1]) X = sm.add_constant(X) res = FirthLogit(y, X).fit() print(res.params) print(res.pvalues) # LRT p-values when pl=True print(res.conf_int()) # profile likelihood CIs when pl=True print(res.summary()) ``` ### Formula interface If you install `firthmodels[formula]` (or `pip install formulaic`), you can fit from a formula and a pandas DataFrame: ```python import pandas as pd from firthmodels.adapters.statsmodels import FirthLogit df = pd.DataFrame({"y": [0, 1, 0, 1], "age": [20, 30, 40, 50], "treatment": [0, 1, 0, 1]}) res = FirthLogit.from_formula("y ~ age + treatment", df).fit() ``` ## API notes ### `FirthLogisticRegression` parameters | Parameter | Default | Description | |-----------|---------|-------------| | `backend` | `'auto'` | `'auto'`, `'numba'`, or `'numpy'`. `'auto'` uses numba if available. | | `fit_intercept` | `True` | Whether to add an intercept term | | `max_iter` | `25` | Maximum Newton-Raphson iterations | | `gtol` | `1e-4` | Gradient convergence tolerance (converged when max\|gradient\| < gtol) | | `xtol` | `1e-4` | Parameter convergence tolerance (converged when max\|delta\| < xtol) | | `max_step` | `5.0` | Maximum step size per coefficient | | `max_halfstep` | `25` | Maximum step-halvings per iteration | | `penalty_weight` | `0.5` | Weight of the Firth penalty term. The default 0.5 corresponds to the standard Firth bias reduction method (Firth, 1993), equivalent to using Jeffreys' invariant prior. Set to `0.0` for unpenalized maximum likelihood estimation. | ### `FirthLogisticRegression` attributes (after fitting) | Attribute | Description | |-----------|-------------| | `coef_` | Coefficient estimates | | `intercept_` | Intercept (0.0 if `fit_intercept=False`) | | `bse_` | Wald standard errors; includes intercept if `fit_intercept=True` | | `pvalues_` | Wald p-values; includes intercept if `fit_intercept=True` | | `loglik_` | Penalized log-likelihood | | `n_iter_` | Number of iterations | | `converged_` | Whether the solver converged | | `lrt_pvalues_` | LRT p-values (after calling `lrt()`); includes intercept if `fit_intercept=True` | | `lrt_bse_` | Back-corrected SEs (after calling `lrt()`); includes intercept if `fit_intercept=True` | | `classes_` | Class labels (shape `(2,)`) | | `n_features_in_` | Number of features seen during fit | | `feature_names_in_` | Feature names (if X had string column names) | ### `FirthCoxPH` parameters | Parameter | Default | Description | |-----------|---------|-------------| | `backend` | `'auto'` | `'auto'`, `'numba'`, or `'numpy'`. `'auto'` uses numba if available. | | `max_iter` | `50` | Maximum Newton-Raphson iterations | | `gtol` | `1e-4` | Gradient convergence tolerance (converged when max\|gradient\| < gtol) | | `xtol` | `1e-6` | Parameter convergence tolerance (converged when max\|delta\| < xtol) | | `max_step` | `5.0` | Maximum step size per coefficient | | `max_halfstep` | `5` | Maximum step-halvings per iteration | | `penalty_weight` | `0.5` | Weight of the Firth penalty term. The default 0.5 corresponds to the standard Firth bias reduction method (Heinze and Schemper, 2001), equivalent to using Jeffreys' invariant prior. Set to `0.0` for unpenalized Cox partial likelihood estimation. | ### `FirthCoxPH` attributes (after fitting) | Attribute | Description | |-----------|-------------| | `coef_` | Coefficient estimates (log hazard ratios) | | `bse_` | Wald standard errors | | `pvalues_` | Wald p-values | | `loglik_` | Penalized log partial likelihood | | `n_iter_` | Number of iterations | | `converged_` | Whether the solver converged | | `lrt_pvalues_` | LRT p-values (after calling `lrt()`) | | `lrt_bse_` | Back-corrected SEs (after calling `lrt()`) | | `unique_times_` | Unique event times (ascending order) | | `cum_baseline_hazard_` | Breslow cumulative baseline hazard at `unique_times_` | | `baseline_survival_` | Baseline survival function at `unique_times_` | | `n_features_in_` | Number of features seen during fit | | `feature_names_in_` | Feature names (if X had string column names) | `predict(X)` returns the linear predictor `X @ coef_` (log partial hazard). `predict_cumulative_hazard_function(X)` and `predict_survival_function(X)` return arrays evaluated at `unique_times_`. ## References Firth D (1993). Bias reduction of maximum likelihood estimates. *Biometrika* 80, 27-38. Heinze G, Schemper M (2001). A solution to the problem of monotone likelihood in Cox regression. *Biometrics* 57, 114-119. Heinze G, Schemper M (2002). A solution to the problem of separation in logistic regression. *Statistics in Medicine* 21, 2409-2419. Konis, K. (2007). Linear Programming Algorithms for Detecting Separated Data in Binary Logistic Regression Models. DPhil thesis, University of Oxford. Kosmidis I, Firth D (2021). Jeffreys-prior penalty, finiteness and shrinkage in binomial-response generalized linear models. *Biometrika* 108, 71-82. Kosmidis I, Schumacher D, Schwendinger F (2022). _detectseparation: Detect and Check for Separation and Infinite Maximum Likelihood Estimates_. doi:10.32614/CRAN.package.detectseparation <https://doi.org/10.32614/CRAN.package.detectseparation>, R package version 0.3, <https://CRAN.R-project.org/package=detectseparation>. Mbatchou J et al. (2021). Computationally efficient whole-genome regression for quantitative and binary traits. *Nature Genetics* 53, 1097-1103. Venzon DJ, Moolgavkar SH (1988). A method for computing profile-likelihood-based confidence intervals. *Applied Statistics* 37, 87-94. ## License MIT
text/markdown
null
Jon Luo <20971593+jzluo@users.noreply.github.com>
null
null
null
bias reduction, cox, cox proportional hazards, firth, firth logistic regression, logistic regression, penalized likelihood, rare events, separation, survival analysis
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Science/Research", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Topic :: Scientific/Engineering" ]
[]
null
null
>=3.10
[]
[]
[]
[ "numpy>=1.24", "scikit-learn>=1.6", "scipy>=1.12", "typing-extensions>=4.15.0; python_version < \"3.11\"", "formulaic>=1.2.1; extra == \"dev\"", "mypy>=1.19.0; extra == \"dev\"", "numba>=0.64; extra == \"dev\"", "pandas>=1.5.3; extra == \"dev\"", "polars>=1.37.1; extra == \"dev\"", "pytest-cov>=7.0.0; extra == \"dev\"", "pytest>=9.0.1; extra == \"dev\"", "ruff>=0.14.8; extra == \"dev\"", "formulaic>=1.2.1; extra == \"formula\"", "numba>=0.64; extra == \"numba\"" ]
[]
[]
[]
[ "Homepage, https://github.com/jzluo/firthmodels", "Repository, https://github.com/jzluo/firthmodels", "Issues, https://github.com/jzluo/firthmodels/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:46:26.194719
firthmodels-0.7.2.tar.gz
636,835
65/ec/406a5939ac57f17d5b278f34a495c4a318383b1352dee2d85230188bba53/firthmodels-0.7.2.tar.gz
source
sdist
null
false
bcc45a587166ae044b97d19586a2585f
28e9b06b2698a6ca14f217ab56d7af2e405744c657eeaa55db42e824ba5ae7fa
65ec406a5939ac57f17d5b278f34a495c4a318383b1352dee2d85230188bba53
MIT
[ "LICENSE" ]
237
2.4
pymqrest
1.1.9
Python wrapper for the IBM MQ REST API
# pymqrest Python wrapper for the IBM MQ administrative REST API. `pymqrest` provides typed Python methods for every MQSC command exposed by the IBM MQ 9.4 `runCommandJSON` REST endpoint. Attribute names are automatically translated between Python `snake_case` and native MQSC parameter names, so you work with Pythonic identifiers throughout. ## Table of Contents - [Installation](#installation) - [Quick start](#quick-start) - [API overview](#api-overview) - [Documentation](#documentation) - [Development](#development) - [License](#license) ## Installation ```bash pip install pymqrest ``` Requires Python 3.12+. ## Quick start ```python from pymqrest import MQRESTSession, LTPAAuth session = MQRESTSession( rest_base_url="https://localhost:9443/ibmmq/rest/v2", qmgr_name="QM1", credentials=LTPAAuth("mqadmin", "mqadmin"), verify_tls=False, ) # Query the queue manager qmgr = session.display_qmgr() print(qmgr["queue_manager_name"]) # List all local queues queues = session.display_qlocal(name="*") for q in queues: print(q["queue_name"], q.get("current_queue_depth", 0)) # Idempotent object management result = session.ensure_qlocal( name="APP.REQUESTS", request_parameters={"max_queue_depth": "50000"}, ) print(result.action) # EnsureAction.CREATED, UPDATED, or UNCHANGED ``` ## API overview ### Session `MQRESTSession` manages authentication, connection settings, and attribute mapping. All command methods are called directly on the session object. ```python MQRESTSession( rest_base_url="https://host:9443/ibmmq/rest/v2", qmgr_name="QM1", credentials=LTPAAuth("user", "pass"), # or CertificateAuth / BasicAuth map_attributes=True, # snake_case <-> MQSC translation (default) mapping_strict=True, # raise on unknown attributes (default) verify_tls=True, # TLS certificate verification (default) timeout_seconds=30.0, # HTTP request timeout (default) ) ``` ### Commands Over 130 generated methods cover the MQSC command set: | Verb | Methods | Returns | Example | | --- | --- | --- | --- | | `display_*` | 44 | `list[dict]` or `dict \| None` | `session.display_qlocal(name="*")` | | `define_*` | 19 | `None` | `session.define_qlocal(name="Q1")` | | `alter_*` | 17 | `None` | `session.alter_qlocal(name="Q1", ...)` | | `delete_*` | 16 | `None` | `session.delete_qlocal(name="Q1")` | | Other | 41 | `None` | `start_channel`, `stop_listener`, `clear_qlocal`, ... | All methods accept optional `request_parameters` and `response_parameters` dicts. `DISPLAY` commands default to returning all attributes. ### Ensure methods Idempotent `ensure_*` methods implement a declarative upsert pattern for 15 object types (queues, channels, topics, listeners, and more): - **DEFINE** when the object does not exist - **ALTER** only the attributes that differ - **No-op** when all specified attributes already match Returns an `EnsureResult` whose `action` attribute is `EnsureAction.CREATED`, `UPDATED`, or `UNCHANGED`. When the action is `UPDATED`, `result.changed` contains the attribute names that differed. ### Attribute mapping When `map_attributes=True` (the default), attribute names and values are translated automatically: | Direction | From | To | Example | | --- | --- | --- | --- | | Request | `max_queue_depth` | `MAXDEPTH` | snake_case to MQSC | | Response | `MAXDEPTH` | `max_queue_depth` | MQSC to snake_case | Disable per-session (`map_attributes=False`) or per-call for raw MQSC parameter access. ### Authentication Three credential types are supported: - `CertificateAuth(cert_path, key_path)` — mutual TLS client certificates - `LTPAAuth(username, password)` — LTPA token login (automatic at session creation) - `BasicAuth(username, password)` — HTTP Basic authentication ## Documentation Full documentation: <https://wphillipmoore.github.io/mq-rest-admin-python/> ## Development ```bash uv sync --group dev uv run python3 scripts/dev/validate_local.py ``` ## License GPL-3.0-or-later. See `LICENSE`.
text/markdown
null
Phillip Moore <w.phillip.moore@gmail.com>
null
null
null
ibm, mq, mqsc, rest-api, messaging, queue-manager
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Networking", "Typing :: Typed" ]
[]
null
null
<4.0,>=3.12
[]
[]
[]
[ "requests" ]
[]
[]
[]
[ "Homepage, https://github.com/wphillipmoore/mq-rest-admin-python", "Documentation, https://wphillipmoore.github.io/mq-rest-admin-python/", "Repository, https://github.com/wphillipmoore/mq-rest-admin-python", "Issues, https://github.com/wphillipmoore/mq-rest-admin-python/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:45:52.819468
pymqrest-1.1.9.tar.gz
55,908
3d/5b/3e57a5bc36c8459459943ca4ff0d70912117842b92efe8b945b944b89780/pymqrest-1.1.9.tar.gz
source
sdist
null
false
b6e91b1e10eb09b14013c2324815584e
9ef4a6fed15d6688a47a96e02869717f08409e19a93f39dd1995572581944554
3d5b3e57a5bc36c8459459943ca4ff0d70912117842b92efe8b945b944b89780
GPL-3.0-or-later
[ "LICENSE" ]
210
2.4
ocg
0.4.1
100% openCypher-compliant in-memory graph database — 4 backends, 175+ algorithms, pure Rust
# OCG — OpenCypher Graph **High-performance in-memory graph database with 100% OpenCypher compliance, 4 backends, and 175+ algorithms — pure Rust.** [![PyPI](https://img.shields.io/pypi/v/ocg)](https://pypi.org/project/ocg) [![Python](https://img.shields.io/pypi/pyversions/ocg)](https://pypi.org/project/ocg) [![License](https://img.shields.io/badge/License-Apache%202.0-blue.svg)](LICENSE) [![OpenCypher TCK](https://img.shields.io/badge/OpenCypher%20TCK-100%25-brightgreen)](https://opencypher.org) [![Rust](https://img.shields.io/badge/rust-1.70+-orange.svg)](https://www.rust-lang.org) ## Overview OCG executes [OpenCypher](https://opencypher.org) queries against in-memory property graphs. It is built in pure Rust and exposed to Python via PyO3 bindings. - **100% OpenCypher TCK**: 3,897 / 3,897 scenarios passing (0 skipped, 0 failed) - **4 graph backends**: PropertyGraph, NetworKitRust, RustworkxCore, Graphrs - **175+ graph algorithms**: centrality, community, pathfinding, spanning trees, flow, coloring, matching, cliques, layout, generators - **Bulk Loader API**: 57x faster batch construction vs OpenCypher `CREATE` statements - **Serialization**: save/load graphs to JSON with full metadata - **Python 3.11–3.14**, macOS · Linux (glibc + musl) · Windows --- ## Installation ```bash pip install ocg ``` Rust: ```toml [dependencies] ocg = "0.4.1" ``` --- ## Quick Start ### Python ```python from ocg import Graph graph = Graph() # OpenCypher queries graph.execute("CREATE (a:Person {name: 'Alice', age: 30})") graph.execute("CREATE (b:Person {name: 'Bob', age: 25})") graph.execute("MATCH (a:Person), (b:Person) WHERE a.name='Alice' AND b.name='Bob' CREATE (a)-[:KNOWS]->(b)") result = graph.execute("MATCH (a:Person)-[:KNOWS]->(b:Person) RETURN a.name AS from, b.name AS to") print(result) # [{'from': 'Alice', 'to': 'Bob'}] ``` ### Bulk Loader (10–57x faster) Bypasses the OpenCypher parser for large batch operations: ```python from ocg import Graph graph = Graph() node_ids = graph.bulk_create_nodes([ (["Person"], {"name": "Alice", "age": 30}), (["Person"], {"name": "Bob", "age": 25}), ]) graph.bulk_create_relationships([ (node_ids[0], node_ids[1], "KNOWS", {"since": 2020}), ]) result = graph.execute("MATCH (a)-[:KNOWS]->(b) RETURN a.name, b.name") ``` ### Serialization ```python graph.save("my_graph.json") loaded = Graph.load("my_graph.json") ``` ### Rust ```rust use ocg::{PropertyGraph, execute}; let mut graph = PropertyGraph::new(); execute(&mut graph, "CREATE (a:Person {name: 'Alice'})").unwrap(); let result = execute(&mut graph, "MATCH (n:Person) RETURN n.name").unwrap(); ``` --- ## Graph Backends | Backend | Class | Description | |---------|-------|-------------| | PropertyGraph | `Graph` | Native petgraph-based property graph | | NetworKitRust | `NetworKitGraph` | Port of NetworKit algorithms to pure Rust | | RustworkxCore | `RustworkxGraph` | IBM Qiskit rustworkx-core algorithms | | Graphrs | `GraphrsGraph` | graphrs-based community detection | All four backends expose identical APIs: OpenCypher execution, bulk loader, 175+ algorithms, and save/load. --- ## Graph Algorithms (175+) All algorithms are available on all 4 backends. | Category | Algorithms | |----------|-----------| | Centrality | degree, betweenness, closeness, pagerank, eigenvector, katz, harmonic, voterank | | Pathfinding | bfs, dijkstra, astar, bellman_ford, floyd_warshall, all_pairs, all_simple_paths, all_pairs_all_simple_paths | | Shortest Paths | single_source, multi_source, k_shortest, average_shortest_path_length | | Spanning Trees | minimum, maximum, steiner_tree | | DAG | topological_sort, is_dag, find_cycles, dag_longest_path, transitive_closure, transitive_reduction, dag_to_tree | | Flow | max_flow, min_cut_capacity | | Coloring | node_coloring, edge_coloring, chromatic_number | | Matching | max_weight_matching, max_cardinality_matching | | Community | louvain, label_propagation, girvan_newman | | Components | connected_components, strongly_connected, number_weakly_connected, is_connected, is_tree, is_forest | | Cliques | find_cliques, max_clique, clique_number, node_clique_number, cliques_containing_node | | Traversal | dfs, bfs_layers, descendants, ancestors | | Transitivity | triangles, transitivity, clustering, average_clustering, square_clustering | | Graph Ops | complement, line_graph, cartesian_product, tensor_product, strong_product, lexicographic_product, graph_power | | Euler | is_eulerian, eulerian_circuit, semieulerian | | Planar | is_planar | | Contraction | contract_nodes, quotient_graph | | Token Swapper | token_swapper | | Generators | erdos_renyi, barabasi_albert, complete_graph, path_graph, cycle_graph, star_graph, grid_graph, petersen_graph, watts_strogatz, configuration_model, expected_degree_graph | | Layout | spring, kamada_kawai, spectral, sfdp, hierarchical, bipartite, circular, shell, random | ```python from ocg import Graph graph = Graph() # ... populate graph ... scores = graph.pagerank(damping=0.85, max_iter=100) communities = graph.louvain() path = graph.dijkstra(source_id, target_id) ``` --- ## Supported OpenCypher Features ### Clauses - `MATCH`, `OPTIONAL MATCH`, variable-length paths `[*1..3]` - `CREATE`, `MERGE`, `SET`, `DELETE`, `DETACH DELETE`, `REMOVE` - `WITH`, `UNWIND`, `RETURN`, `WHERE` - `ORDER BY`, `SKIP`, `LIMIT`, `DISTINCT` - `UNION`, `UNION ALL` ### Expressions - Property access, list indexing, string slicing - Arithmetic: `+`, `-`, `*`, `/`, `%`, `^` - Comparison: `=`, `<>`, `<`, `>`, `<=`, `>=` - Logical: `AND`, `OR`, `NOT`, `XOR` - String: `STARTS WITH`, `ENDS WITH`, `CONTAINS`, `=~` - Null: `IS NULL`, `IS NOT NULL` - List: `IN`, comprehensions, quantifiers ### Functions (60+) - **String**: `substring`, `trim`, `toLower`, `toUpper`, `split`, `replace` - **Math**: `abs`, `ceil`, `floor`, `round`, `sqrt`, `sin`, `cos`, `log` - **List**: `size`, `head`, `tail`, `range`, `reverse`, `keys` - **Aggregation**: `count`, `sum`, `avg`, `min`, `max`, `collect` - **Temporal**: `date`, `datetime`, `localDatetime`, `duration` - **Predicates**: `exists`, `all`, `any`, `none`, `single` ### Procedures - `db.labels()`, `db.relationshipTypes()`, `db.propertyKeys()` - `dbms.components()` --- ## TCK Compliance **3,897 / 3,897 scenarios passing — 100% (0 skipped, 0 failed)** Validated against the [openCypher Technology Compatibility Kit](https://github.com/opencypher/openCypher). --- ## Development ```bash # Build cargo build --release # Unit tests cargo test --no-default-features # OpenCypher TCK cargo test --test tck_property_graph --no-default-features # Python wheel (requires maturin) maturin develop --features python ``` --- ## Credits - **[petgraph](https://github.com/petgraph/petgraph)** — core graph data structures (MIT/Apache-2.0) - **[rustworkx-core](https://github.com/Qiskit/rustworkx)** — graph algorithms (Apache-2.0) - **[graphrs](https://github.com/malcolmvr/graphrs)** — community detection (MIT) - **[openCypher TCK](https://github.com/opencypher/openCypher)** — test compatibility kit (Apache-2.0) - Algorithm designs inspired by [NetworKit](https://networkit.github.io/) (MIT) Algorithm implementations (PageRank, Betweenness Centrality, Dijkstra, etc.) are based on published academic work. See NOTICE file for complete citations. --- ## License Apache-2.0 — see [LICENSE](LICENSE) and [NOTICE](NOTICE) files. OpenCypher® and Cypher® are registered trademarks of Neo4j, Inc. This project implements the open [OpenCypher specification](https://opencypher.org) and is not affiliated with Neo4j. --- ## Contributing Issues and proposals may be submitted via GitHub. Contributions are evaluated on a controlled schedule — pull requests are reviewed at the maintainer's discretion and timeline.
text/markdown; charset=UTF-8; variant=GFM
null
Gregorio Momm <gregoriomomm@gmail.com>
null
null
Apache-2.0
graph, database, cypher, opencypher, graph-database, query-language, rust, graph-algorithms, bulk-loader
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Rust", "Topic :: Database", "Topic :: Database :: Database Engines/Servers", "Topic :: Scientific/Engineering :: Mathematics", "Topic :: Software Development :: Libraries :: Python Modules" ]
[]
null
null
>=3.8
[]
[]
[]
[]
[]
[]
[]
[ "Documentation, https://github.ibm.com/enjoycode/ocg", "Homepage, https://github.ibm.com/enjoycode/ocg", "Repository, https://github.ibm.com/enjoycode/ocg" ]
twine/6.2.0 CPython/3.14.3
2026-02-20T21:45:44.282100
ocg-0.4.1.tar.gz
513,431
c5/e0/7a7c2298df39d93357fd132dce77729e15cd09f9a67a5126218ab54d4a75/ocg-0.4.1.tar.gz
source
sdist
null
false
d730be8eb72d2e06dea754ba7af774c2
ec9204f89eead810771fc448cc3bc35e419a26d4bff6e25544a2fd6dc13ef074
c5e07a7c2298df39d93357fd132dce77729e15cd09f9a67a5126218ab54d4a75
null
[ "LICENSE", "NOTICE" ]
1,438
2.4
agentsteer
0.4.1
Runtime protection for AI coding agents
# AgentSteer **Runtime protection for AI coding agents.** [![License: MIT](https://img.shields.io/badge/License-MIT-blue.svg)](https://opensource.org/licenses/MIT) [![Python 3.10+](https://img.shields.io/badge/python-3.10+-blue.svg)](https://www.python.org/downloads/) ## What it does AgentSteer intercepts every tool call your AI agent makes via a PreToolUse hook, scores it against the task description, and blocks unauthorized actions before they execute. It catches prompt injection attacks, unauthorized file access, data exfiltration through delegation, and off-task behavior. Scoring uses [oss-safeguard-20b](https://openrouter.ai/openai/gpt-oss-safeguard-20b) (20B parameter model) via OpenRouter at under 200ms median latency. ## Quick Start ```bash pip install agentsteer ``` Set your OpenRouter API key. You can use a `.env` file or export it directly: ```bash export OPENROUTER_API_KEY=sk-or-... ``` Install the hook for your framework: ```bash agentsteer install claude-code ``` That's it. Every tool call is now scored and unauthorized actions are blocked. ## How It Works ``` Agent makes tool call | v PreToolUse hook intercepts the call | v AgentSteer scores the action against the task description (oss-safeguard-20b via OpenRouter, <200ms median) | v Score >= threshold? ---> BLOCK (action never executes) | v Score < threshold? ---> ALLOW (action proceeds normally) ``` Each action is evaluated against 4 security policies. If any policy is violated, the action is blocked and the agent receives feedback explaining why. ## Supported Frameworks | Framework | Install command | Hook type | |-----------|----------------|-----------| | [Claude Code](https://github.com/anthropics/claude-code) | `agentsteer install claude-code` | PreToolUse hook via `settings.json` | | [OpenHands](https://github.com/All-Hands-AI/OpenHands) | `agentsteer install openhands` | EventStream middleware | | Python API | `from agentsteer import score_action` | Direct function call | ## Results Evaluated on the [AgentDojo](https://github.com/ethz-spylab/agentdojo) benchmark, which tests whether agents can be tricked into executing prompt injection attacks hidden in their environment. | Framework | Condition | Attacks blocked | Utility preserved | n | |-----------|-----------|----------------|-------------------|---| | Claude Code | Honest tasks | 0% false positives | 100% | 20 | | Claude Code | Attack tasks | 100% blocked | -- | 20 | | OpenHands | Honest tasks | 0% false positives | 100% | 20 | | OpenHands | Attack tasks | 95% blocked | -- | 20 | | OpenHands | Per-action detection | 96.5% detected | 0% false positives | 1000 | Median scoring latency: <200ms. Cost: $0.38 per 1M tokens via OpenRouter. Full evaluation results and methodology at [agentsteer.ai/evaluations](https://agentsteer.ai/evaluations). ## Security Policies | Policy | What it catches | Example | |--------|----------------|---------| | P1: Read-only | Write actions when the task only requires reading | Task says "look up calendar" but agent sends an email | | P2: Delegation | Actions that hand off control or leak data to external parties | Agent posts task contents to a URL, sends to unrelated email | | P3: Category mismatch | Actions in a different category than the task | Task is about calendar but agent modifies files | | P4: Target mismatch | Actions targeting a different entity than specified | Task says "email Alice" but agent emails Bob | A post-filter suppresses false positives from self-correction patterns (e.g., agent cancels a calendar event it just created in error). ## Configuration | Environment variable | Description | Default | |---------------------|-------------|---------| | `OPENROUTER_API_KEY` | OpenRouter API key for scoring model | required | | `AGENT_STEER_TASK` | Task description to score actions against | read from agent context | | `AGENT_STEER_THRESHOLD` | Score threshold for blocking (0.0 to 1.0) | `0.80` | | `AGENT_STEER_DEBUG` | Enable debug logging (`1` or `true`) | off | You can also run `agentsteer setup` for an interactive configuration wizard, or `agentsteer login` to connect to the AgentSteer cloud dashboard for real-time monitoring. ## CLI Reference ```bash agentsteer setup # Interactive first-run setup agentsteer install claude-code # Install hook for Claude Code agentsteer status # Show current configuration agentsteer score "task" "action" # Score a single action agentsteer sessions # List monitored sessions agentsteer version # Print version ``` ## Links - **Website**: [agentsteer.ai](https://agentsteer.ai) - **Documentation**: [agentsteer.ai/docs](https://agentsteer.ai/docs) - **Evaluations**: [agentsteer.ai/evaluations](https://agentsteer.ai/evaluations) - **Repository**: [github.com/AgentSteer/AgentSteer](https://github.com/AgentSteer/AgentSteer) ## License MIT
text/markdown
Ram Rachum
null
null
null
MIT
agent, agentsteer, ai, llm, monitor, security
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Topic :: Security" ]
[]
null
null
>=3.10
[]
[]
[]
[ "python-dotenv>=1.0", "requests>=2.28", "boto3; extra == \"benchmarks\"", "plotly; extra == \"benchmarks\"" ]
[]
[]
[]
[ "Homepage, https://agentsteer.ai", "Documentation, https://agentsteer.ai/docs/", "Repository, https://github.com/AgentSteer/AgentSteer" ]
twine/6.2.0 CPython/3.12.3
2026-02-20T21:45:33.520769
agentsteer-0.4.1.tar.gz
28,950
97/36/05af9e60b3017a3a6b3afb1611d411b7b0b3aa3621edd52a831f1cc220d6/agentsteer-0.4.1.tar.gz
source
sdist
null
false
3923ac91b12bc44d4426f9c1354dc868
0624dfd429c9ec9bb21fdffd5d2e87691699e33c28ff315b2536d8d1c1e2275b
973605af9e60b3017a3a6b3afb1611d411b7b0b3aa3621edd52a831f1cc220d6
null
[]
216
2.4
pyvex
9.2.202
A Python interface to libVEX and VEX IR
# PyVEX [![Latest Release](https://img.shields.io/pypi/v/pyvex.svg)](https://pypi.python.org/pypi/pyvex/) [![Python Version](https://img.shields.io/pypi/pyversions/pyvex)](https://pypi.python.org/pypi/pyvex/) [![PyPI Statistics](https://img.shields.io/pypi/dm/pyvex.svg)](https://pypistats.org/packages/pyvex) [![License](https://img.shields.io/github/license/angr/pyvex.svg)](https://github.com/angr/pyvex/blob/master/LICENSE) PyVEX is Python bindings for the VEX IR. ## Project Links Project repository: https://github.com/angr/pyvex Documentation: https://api.angr.io/projects/pyvex/en/latest/ ## Installing PyVEX PyVEX can be pip-installed: ```bash pip install pyvex ``` ## Using PyVEX ```python import pyvex import archinfo # translate an AMD64 basic block (of nops) at 0x400400 into VEX irsb = pyvex.lift(b"\x90\x90\x90\x90\x90", 0x400400, archinfo.ArchAMD64()) # pretty-print the basic block irsb.pp() # this is the IR Expression of the jump target of the unconditional exit at the end of the basic block print(irsb.next) # this is the type of the unconditional exit (i.e., a call, ret, syscall, etc) print(irsb.jumpkind) # you can also pretty-print it irsb.next.pp() # iterate through each statement and print all the statements for stmt in irsb.statements: stmt.pp() # pretty-print the IR expression representing the data, and the *type* of that IR expression written by every store statement import pyvex for stmt in irsb.statements: if isinstance(stmt, pyvex.IRStmt.Store): print("Data:", end="") stmt.data.pp() print("") print("Type:", end="") print(stmt.data.result_type) print("") # pretty-print the condition and jump target of every conditional exit from the basic block for stmt in irsb.statements: if isinstance(stmt, pyvex.IRStmt.Exit): print("Condition:", end="") stmt.guard.pp() print("") print("Target:", end="") stmt.dst.pp() print("") # these are the types of every temp in the IRSB print(irsb.tyenv.types) # here is one way to get the type of temp 0 print(irsb.tyenv.types[0]) ``` Keep in mind that this is a *syntactic* representation of a basic block. That is, it'll tell you what the block means, but you don't have any context to say, for example, what *actual* data is written by a store instruction. ## VEX Intermediate Representation To deal with widely diverse architectures, it is useful to carry out analyses on an intermediate representation. An IR abstracts away several architecture differences when dealing with different architectures, allowing a single analysis to be run on all of them: - **Register names.** The quantity and names of registers differ between architectures, but modern CPU designs hold to a common theme: each CPU contains several general purpose registers, a register to hold the stack pointer, a set of registers to store condition flags, and so forth. The IR provides a consistent, abstracted interface to registers on different platforms. Specifically, VEX models the registers as a separate memory space, with integer offsets (i.e., AMD64's `rax` is stored starting at address 16 in this memory space). - **Memory access.** Different architectures access memory in different ways. For example, ARM can access memory in both little-endian and big-endian modes. The IR must abstract away these differences. - **Memory segmentation.** Some architectures, such as x86, support memory segmentation through the use of special segment registers. The IR understands such memory access mechanisms. - **Instruction side-effects.** Most instructions have side-effects. For example, most operations in Thumb mode on ARM update the condition flags, and stack push/pop instructions update the stack pointer. Tracking these side-effects in an *ad hoc* manner in the analysis would be crazy, so the IR makes these effects explicit. There are lots of choices for an IR. We use VEX, since the uplifting of binary code into VEX is quite well supported. VEX is an architecture-agnostic, side-effects-free representation of a number of target machine languages. It abstracts machine code into a representation designed to make program analysis easier. This representation has five main classes of objects: - **Expressions.** IR Expressions represent a calculated or constant value. This includes memory loads, register reads, and results of arithmetic operations. - **Operations.** IR Operations describe a *modification* of IR Expressions. This includes integer arithmetic, floating-point arithmetic, bit operations, and so forth. An IR Operation applied to IR Expressions yields an IR Expression as a result. - **Temporary variables.** VEX uses temporary variables as internal registers: IR Expressions are stored in temporary variables between use. The content of a temporary variable can be retrieved using an IR Expression. These temporaries are numbered, starting at `t0`. These temporaries are strongly typed (i.e., "64-bit integer" or "32-bit float"). - **Statements.** IR Statements model changes in the state of the target machine, such as the effect of memory stores and register writes. IR Statements use IR Expressions for values they may need. For example, a memory store *IR Statement* uses an *IR Expression* for the target address of the write, and another *IR Expression* for the content. - **Blocks.** An IR Block is a collection of IR Statements, representing an extended basic block (termed "IR Super Block" or "IRSB") in the target architecture. A block can have several exits. For conditional exits from the middle of a basic block, a special *Exit* IR Statement is used. An IR Expression is used to represent the target of the unconditional exit at the end of the block. VEX IR is actually quite well documented in the `libvex_ir.h` file (https://github.com/angr/vex/blob/dev/pub/libvex_ir.h) in the VEX repository. For the lazy, we'll detail some parts of VEX that you'll likely interact with fairly frequently. To begin with, here are some IR Expressions: | IR Expression | Evaluated Value | VEX Output Example | | ------------- | --------------- | ------- | | Constant | A constant value. | 0x4:I32 | | Read Temp | The value stored in a VEX temporary variable. | RdTmp(t10) | | Get Register | The value stored in a register. | GET:I32(16) | | Load Memory | The value stored at a memory address, with the address specified by another IR Expression. | LDle:I32 / LDbe:I64 | | Operation | A result of a specified IR Operation, applied to specified IR Expression arguments. | Add32 | | If-Then-Else | If a given IR Expression evaluates to 0, return one IR Expression. Otherwise, return another. | ITE | | Helper Function | VEX uses C helper functions for certain operations, such as computing the conditional flags registers of certain architectures. These functions return IR Expressions. | function\_name() | These expressions are then, in turn, used in IR Statements. Here are some common ones: | IR Statement | Meaning | VEX Output Example | | ------------ | ------- | ------------------ | Write Temp | Set a VEX temporary variable to the value of the given IR Expression. | WrTmp(t1) = (IR Expression) | Put Register | Update a register with the value of the given IR Expression. | PUT(16) = (IR Expression) | Store Memory | Update a location in memory, given as an IR Expression, with a value, also given as an IR Expression. | STle(0x1000) = (IR Expression) | Exit | A conditional exit from a basic block, with the jump target specified by an IR Expression. The condition is specified by an IR Expression. | if (condition) goto (Boring) 0x4000A00:I32 | An example of an IR translation, on ARM, is produced below. In the example, the subtraction operation is translated into a single IR block comprising 5 IR Statements, each of which contains at least one IR Expression (although, in real life, an IR block would typically consist of more than one instruction). Register names are translated into numerical indices given to the *GET* Expression and *PUT* Statement. The astute reader will observe that the actual subtraction is modeled by the first 4 IR Statements of the block, and the incrementing of the program counter to point to the next instruction (which, in this case, is located at `0x59FC8`) is modeled by the last statement. The following ARM instruction: subs R2, R2, #8 Becomes this VEX IR: t0 = GET:I32(16) t1 = 0x8:I32 t3 = Sub32(t0,t1) PUT(16) = t3 PUT(68) = 0x59FC8:I32 Cool stuff! ## Citing PyVEX If you use PyVEX in an academic work, please cite the paper for which it was developed: ```bibtex @article{shoshitaishvili2015firmalice, title={Firmalice - Automatic Detection of Authentication Bypass Vulnerabilities in Binary Firmware}, author={Shoshitaishvili, Yan and Wang, Ruoyu and Hauser, Christophe and Kruegel, Christopher and Vigna, Giovanni}, booktitle={NDSS}, year={2015} } ```
text/markdown
null
null
null
null
null
null
[ "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13" ]
[]
null
null
>=3.10
[]
[]
[]
[ "bitstring", "cffi>=1.0.3; implementation_name == \"cpython\"", "furo; extra == \"docs\"", "myst-parser; extra == \"docs\"", "sphinx; extra == \"docs\"", "sphinx-autodoc-typehints; extra == \"docs\"", "atheris>=2.3.0; extra == \"fuzzing\"", "pytest; extra == \"testing\"", "pytest-xdist; extra == \"testing\"" ]
[]
[]
[]
[ "Homepage, https://api.angr.io/projects/pyvex/en/latest/", "Repository, https://github.com/angr/pyvex" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:44:30.915795
pyvex-9.2.202.tar.gz
3,649,074
7c/eb/7fc96c24eb9c7485cf7b8a9796617a61e7a746fc9cc0ebab7fc785170ee6/pyvex-9.2.202.tar.gz
source
sdist
null
false
e8da7e2039655f4167bc7638d1a4d849
0838534da7f4cfbd239966ea38b173d1b0539615ad7c0c0adb82194e7be46fe3
7ceb7fc96c24eb9c7485cf7b8a9796617a61e7a746fc9cc0ebab7fc785170ee6
BSD-2-Clause AND GPL-2.0-only
[ "LICENSE", "pyvex_c/LICENSE", "vex/LICENSE.GPL", "vex/LICENSE.README" ]
2,877
2.4
cle
9.2.202
CLE Loads Everything (at least, many binary formats!) and provides a pythonic interface to analyze what they are and what they would look like in memory.
# CLE [![Latest Release](https://img.shields.io/pypi/v/cle.svg)](https://pypi.python.org/pypi/cle/) [![Python Version](https://img.shields.io/pypi/pyversions/cle)](https://pypi.python.org/pypi/cle/) [![PyPI Statistics](https://img.shields.io/pypi/dm/cle.svg)](https://pypistats.org/packages/cle) [![License](https://img.shields.io/github/license/angr/cle.svg)](https://github.com/angr/cle/blob/master/LICENSE) CLE loads binaries and their associated libraries, resolves imports and provides an abstraction of process memory the same way as if it was loader by the OS's loader. ## Project Links Project repository: https://github.com/angr/cle Documentation: https://api.angr.io/projects/cle/en/latest/ ## Installation ```sh pip install cle ``` ## Usage example ```python >>> import cle >>> ld = cle.Loader("/bin/ls") >>> hex(ld.main_object.entry) '0x4048d0' >>> ld.shared_objects {'ld-linux-x86-64.so.2': <ELF Object ld-2.21.so, maps [0x5000000:0x522312f]>, 'libacl.so.1': <ELF Object libacl.so.1.1.0, maps [0x2000000:0x220829f]>, 'libattr.so.1': <ELF Object libattr.so.1.1.0, maps [0x4000000:0x4204177]>, 'libc.so.6': <ELF Object libc-2.21.so, maps [0x3000000:0x33a1a0f]>, 'libcap.so.2': <ELF Object libcap.so.2.24, maps [0x1000000:0x1203c37]>} >>> ld.addr_belongs_to_object(0x5000000) <ELF Object ld-2.21.so, maps [0x5000000:0x522312f]> >>> libc_main_reloc = ld.main_object.imports['__libc_start_main'] >>> hex(libc_main_reloc.addr) # Address of GOT entry for libc_start_main '0x61c1c0' >>> import pyvex >>> some_text_data = ld.memory.load(ld.main_object.entry, 0x100) >>> irsb = pyvex.lift(some_text_data, ld.main_object.entry, ld.main_object.arch) >>> irsb.pp() IRSB { t0:Ity_I32 t1:Ity_I32 t2:Ity_I32 t3:Ity_I64 t4:Ity_I64 t5:Ity_I64 t6:Ity_I32 t7:Ity_I64 t8:Ity_I32 t9:Ity_I64 t10:Ity_I64 t11:Ity_I64 t12:Ity_I64 t13:Ity_I64 t14:Ity_I64 15 | ------ IMark(0x4048d0, 2, 0) ------ 16 | t5 = 32Uto64(0x00000000) 17 | PUT(rbp) = t5 18 | t7 = GET:I64(rbp) 19 | t6 = 64to32(t7) 20 | t2 = t6 21 | t9 = GET:I64(rbp) 22 | t8 = 64to32(t9) 23 | t1 = t8 24 | t0 = Xor32(t2,t1) 25 | PUT(cc_op) = 0x0000000000000013 26 | t10 = 32Uto64(t0) 27 | PUT(cc_dep1) = t10 28 | PUT(cc_dep2) = 0x0000000000000000 29 | t11 = 32Uto64(t0) 30 | PUT(rbp) = t11 31 | PUT(rip) = 0x00000000004048d2 32 | ------ IMark(0x4048d2, 3, 0) ------ 33 | t12 = GET:I64(rdx) 34 | PUT(r9) = t12 35 | PUT(rip) = 0x00000000004048d5 36 | ------ IMark(0x4048d5, 1, 0) ------ 37 | t4 = GET:I64(rsp) 38 | t3 = LDle:I64(t4) 39 | t13 = Add64(t4,0x0000000000000008) 40 | PUT(rsp) = t13 41 | PUT(rsi) = t3 42 | PUT(rip) = 0x00000000004048d6 43 | t14 = GET:I64(rip) NEXT: PUT(rip) = t14; Ijk_Boring } ``` ## Valid options For a full listing and description of the options that can be provided to the loader and the methods it provides, please examine the docstrings in `cle/loader.py`. If anything is unclear or poorly documented (there is much) please complain through whatever channel you feel appropriate. ## Loading Backends CLE's loader is implemented in the Loader class. There are several backends that can be used to load a single file: - ELF, as its name says, loads ELF binaries. ELF files loaded this way are statically parsed using PyElfTools. - PE is a backend to load Microsoft's Portable Executable format, effectively Windows binaries. It uses the (optional) `pefile` module. - Mach-O is a backend to load, you guessed it, Mach-O binaries. Support is limited for this backend. - Blob is a backend to load unknown data. It requires that you specify the architecture it would be run on, in the form of a class from ArchInfo. Which backend you use can be specified as an argument to Loader. If left unspecified, the loader will pick a reasonable default. ## Finding shared libraries - If the `auto_load_libs` option is set to False, the Loader will not automatically load libraries requested by loaded objects. Otherwise... - The loader determines which shared objects are needed when loading binaries, and searches for them in the following order: - in the current working directory - in folders specified in the `ld_path` option - in the same folder as the main binary - in the system (in the corresponding library path for the architecture of the binary, e.g., /usr/arm-linux-gnueabi/lib for ARM, note that you need to install cross libraries for this, e.g., libc6-powerpc-cross on Debian - needs emdebian repos) - in the system, but with mismatched version numbers from what is specified as a dependency, if the `ignore_import_version_numbers` option is True - If no binary is found with the correct architecture, the loader raises an exception if `except_missing_libs` option is True. Otherwise it simply leaves the dependencies unresolved.
text/markdown
null
null
null
null
null
null
[ "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13" ]
[]
null
null
>=3.10
[]
[]
[]
[ "archinfo==9.2.202", "arpy==1.1.1", "cart", "minidump>=0.0.10", "pefile", "pyelftools>=0.29", "pyvex==9.2.202", "pyxbe~=1.0.3", "pyxdia~=0.1", "sortedcontainers>=2.0", "uefi-firmware>=1.10", "furo; extra == \"docs\"", "myst-parser; extra == \"docs\"", "sphinx; extra == \"docs\"", "sphinx-autodoc-typehints; extra == \"docs\"", "pypcode>=1.1; extra == \"pcode\"", "cffi; extra == \"testing\"", "pypcode>=1.1; extra == \"testing\"", "pytest; extra == \"testing\"", "pytest-xdist; extra == \"testing\"", "types-pefile; extra == \"testing\"" ]
[]
[]
[]
[ "Homepage, https://api.angr.io/projects/cle/en/latest/", "Repository, https://github.com/angr/cle" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:44:29.699001
cle-9.2.202.tar.gz
213,401
c3/17/7c6f2965f520bf18825a01fa583fff4fbfd6598e10b87510c32dc5569f29/cle-9.2.202.tar.gz
source
sdist
null
false
04fc069ff2df336173394b88fe9c1f65
e3105de42cb7e38c3f124647305292df645851b3fc79aa0ddd83dd24dcb3caaa
c3177c6f2965f520bf18825a01fa583fff4fbfd6598e10b87510c32dc5569f29
BSD-2-Clause
[ "LICENSE" ]
988
2.4
claripy
9.2.202
An abstraction layer for constraint solvers
# claripy [![Latest Release](https://img.shields.io/pypi/v/claripy.svg)](https://pypi.python.org/pypi/claripy/) [![Python Version](https://img.shields.io/pypi/pyversions/claripy)](https://pypi.python.org/pypi/claripy/) [![PyPI Statistics](https://img.shields.io/pypi/dm/claripy.svg)](https://pypistats.org/packages/claripy) [![License](https://img.shields.io/github/license/angr/claripy.svg)](https://github.com/angr/claripy/blob/master/LICENSE) Claripy is an abstracted constraint-solving wrapper. ## Project Links Project repository: https://github.com/angr/claripy Documentation: https://api.angr.io/projects/claripy/en/latest/ ## Usage It is usable! General usage is similar to Z3: ```python >>> import claripy >>> a = claripy.BVV(3, 32) >>> b = claripy.BVS('var_b', 32) >>> s = claripy.Solver() >>> s.add(b > a) >>> print(s.eval(b, 1)[0]) ```
text/markdown
null
null
null
null
null
null
[ "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12" ]
[]
null
null
>=3.10
[]
[]
[]
[ "cachetools", "typing-extensions", "z3-solver==4.13.0.0", "furo; extra == \"docs\"", "myst-parser; extra == \"docs\"", "sphinx; extra == \"docs\"", "sphinx-autodoc-typehints; extra == \"docs\"" ]
[]
[]
[]
[ "Homepage, https://github.com/angr/claripy" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:44:28.723081
claripy-9.2.202.tar.gz
146,893
d7/14/8c19a7ccf6cb0054de5858f8357e8f70d7a7456187e6969d5f9497411737/claripy-9.2.202.tar.gz
source
sdist
null
false
538e1e33c4c7c061b1fb3b6ceaa08dcf
bbda4a32033fafdca423b3686dfc1336160529a73cdf2671f056ccad2b901abc
d7148c19a7ccf6cb0054de5858f8357e8f70d7a7456187e6969d5f9497411737
BSD-2-Clause
[ "LICENSE" ]
1,009
2.4
ai-reviewbot
1.0.0a8
AI-powered code review agent for CI/CD pipelines
# AI ReviewBot [![PyPI version](https://img.shields.io/pypi/v/ai-reviewbot)](https://pypi.org/project/ai-reviewbot/) [![Python 3.13+](https://img.shields.io/badge/python-3.13+-blue.svg)](https://www.python.org/downloads/) [![License](https://img.shields.io/badge/license-Apache%202.0-blue.svg)](LICENSE) [![Tests](https://github.com/KonstZiv/ai-code-reviewer/actions/workflows/tests.yml/badge.svg)](https://github.com/KonstZiv/ai-code-reviewer/actions/workflows/tests.yml) [![codecov](https://codecov.io/gh/KonstZiv/ai-code-reviewer/branch/main/graph/badge.svg)](https://codecov.io/gh/KonstZiv/ai-code-reviewer) AI-powered code review tool for **GitHub** and **GitLab** that provides intelligent feedback with **inline suggestions** and one-click "Apply" button. <p align="center"> <a href="https://konstziv.github.io/ai-code-reviewer/">📚 Documentation</a> • <a href="https://konstziv.github.io/ai-code-reviewer/quick-start/">🚀 Quick Start</a> • <a href="https://github.com/marketplace/actions/ai-code-reviewer">🛒 GitHub Marketplace</a> </p> --- ## ✨ Features - 🤖 **AI-Powered Analysis** — Uses Google Gemini for deep code understanding - 💡 **Inline Suggestions** — Comments directly on code lines with GitHub's "Apply suggestion" button - 🔒 **Security Focus** — Identifies vulnerabilities with severity levels (Critical, Warning, Info) - 🌍 **Multi-Language** — Responds in your PR/MR language (adaptive mode) - ✨ **Good Practices** — Highlights what you're doing right, not just issues - 📊 **Transparent Metrics** — Shows tokens, latency, and estimated cost - 🦊 **GitHub & GitLab** — Native support for both platforms ## 🚀 Quick Start ### GitHub Actions (Recommended) ```yaml # .github/workflows/ai-review.yml name: AI Code Review on: pull_request: types: [opened, synchronize] jobs: review: runs-on: ubuntu-latest permissions: contents: read pull-requests: write steps: - uses: KonstZiv/ai-code-reviewer@v1 with: github_token: ${{ secrets.GITHUB_TOKEN }} google_api_key: ${{ secrets.GOOGLE_API_KEY }} ``` ### GitLab CI ```yaml # .gitlab-ci.yml ai-review: image: ghcr.io/konstziv/ai-code-reviewer:1 script: - ai-review rules: - if: $CI_PIPELINE_SOURCE == "merge_request_event" variables: GOOGLE_API_KEY: $GOOGLE_API_KEY GITLAB_TOKEN: $GITLAB_TOKEN # Project Access Token with 'api' scope ``` ### PyPI ```bash pip install ai-reviewbot # Set environment variables export GOOGLE_API_KEY="your-key" export GITHUB_TOKEN="your-token" # Run review ai-review --repo owner/repo --pr 123 ``` ### Docker ```bash # DockerHub docker pull koszivdocker/ai-reviewbot:1 # GitHub Container Registry docker pull ghcr.io/konstziv/ai-code-reviewer:1 ``` ## 📖 Documentation Full documentation available in **6 languages**: | Language | Link | |----------|------| | 🇬🇧 English | [Documentation](https://konstziv.github.io/ai-code-reviewer/) | | 🇺🇦 Українська | [Документація](https://konstziv.github.io/ai-code-reviewer/uk/) | | 🇩🇪 Deutsch | [Dokumentation](https://konstziv.github.io/ai-code-reviewer/de/) | | 🇪🇸 Español | [Documentación](https://konstziv.github.io/ai-code-reviewer/es/) | | 🇲🇪 Crnogorski | [Dokumentacija](https://konstziv.github.io/ai-code-reviewer/sr/) | | 🇮🇹 Italiano | [Documentazione](https://konstziv.github.io/ai-code-reviewer/it/) | ## ⚙️ Configuration | Variable | Default | Description | |----------|---------|-------------| | `GOOGLE_API_KEY` | — | **Required.** Google Gemini API key | | `GITHUB_TOKEN` | — | GitHub token (for GitHub) | | `GITLAB_TOKEN` | — | GitLab token (for GitLab) | | `LANGUAGE` | `en` | Response language (ISO 639 code) | | `LANGUAGE_MODE` | `adaptive` | `adaptive` (detect from PR) or `fixed` | | `GEMINI_MODEL` | `gemini-2.5-flash` | Gemini model to use | | `LOG_LEVEL` | `INFO` | Logging level | See [Configuration Guide](https://konstziv.github.io/ai-code-reviewer/configuration/) for all options. ## 🎯 Example Output The reviewer provides structured feedback with inline suggestions: ### Summary Comment > **🤖 AI Code Review** > > **📊 Summary** — Found 2 issues and 1 good practice. > > | Category | Critical | Warning | Info | > |----------|----------|---------|------| > | Security | 1 | 0 | 0 | > | Code Quality | 0 | 1 | 0 | > > **✨ Good Practices** — Excellent error handling in `api/handlers.py` > > --- > ⏱️ 1.2s | 🪙 1,540 tokens | 💰 ~$0.002 ### Inline Comment with "Apply" Button > ⚠️ **SQL Injection Risk** > > User input is concatenated directly into SQL query. > > ```suggestion > cursor.execute("SELECT * FROM users WHERE id = ?", (user_id,)) > ``` > > 💡 **Why this matters:** SQL injection allows attackers to execute arbitrary SQL commands. Always use parameterized queries. > > 📚 [Learn more](https://owasp.org/www-community/attacks/SQL_Injection) ## 🛠️ Development ```bash # Clone repository git clone https://github.com/KonstZiv/ai-code-reviewer.git cd ai-code-reviewer # Install dependencies with uv uv sync --all-groups # Run tests uv run pytest # Run linters uv run ruff check . uv run mypy src/ # Build documentation uv run mkdocs serve ``` ## 📦 Installation Options | Method | Command | Best For | |--------|---------|----------| | **GitHub Action** | `uses: KonstZiv/ai-code-reviewer@v1` | GitHub projects | | **Docker** | `docker pull koszivdocker/ai-reviewbot` | GitLab CI | | **PyPI** | `pip install ai-reviewbot` | Local testing | ## 💰 Cost Estimate Using Gemini 2.5 Flash: - **Input:** $0.075 / 1M tokens - **Output:** $0.30 / 1M tokens - **Average review:** ~$0.002 (1,500 tokens) 100 reviews/month ≈ **$0.20** ## 📄 License Apache 2.0 — See [LICENSE](LICENSE) for details. ## 🤝 Contributing Contributions are welcome! See [Contributing Guide](CONTRIBUTING.md). ## 📬 Support - 🐛 [Report a Bug](https://github.com/KonstZiv/ai-code-reviewer/issues/new?template=bug_report.md) - 💡 [Request a Feature](https://github.com/KonstZiv/ai-code-reviewer/issues/new?template=feature_request.md) - 📚 [Documentation](https://konstziv.github.io/ai-code-reviewer/) --- <p align="center"> Made with ❤️ by <a href="https://github.com/KonstZiv">Kostyantin Zivenko</a> </p>
text/markdown
null
Kostyantin Zivenko <kos.zivenko@gmail.com>
null
null
Apache-2.0
ai, ci-cd, code-review, github, gitlab, llm
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14" ]
[]
null
null
>=3.13
[]
[]
[]
[ "aiofiles>=25.1.0", "anthropic>=0.76.0", "click>=8.3.1", "gitpython>=3.1.46", "google-api-core>=2.29.0", "google-genai>=1.59.0", "httpx>=0.28.1", "langchain-anthropic>=1.3.1", "langchain-deepseek>=1.0.1", "langchain-google-genai>=4.2.0", "langchain-ollama>=1.0.1", "langchain-openai>=1.1.7", "langchain>=1.2.6", "langgraph>=1.0.6", "openai>=2.15.0", "pydantic-settings>=2.12.0", "pydantic>=2.12.5", "pygithub>=2.8.1", "python-dotenv>=1.2.1", "python-gitlab>=7.1.0", "python-iso639>=0.0.10", "pyyaml>=6.0.3", "rich>=14.2.0", "tenacity>=9.0.0", "typer>=0.21.1" ]
[]
[]
[]
[ "Homepage, https://github.com/KonstZiv/ai-code-reviewer", "Documentation, https://konstziv.github.io/ai-code-reviewer", "Repository, https://github.com/KonstZiv/ai-code-reviewer", "Issues, https://github.com/KonstZiv/ai-code-reviewer/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:44:27.874492
ai_reviewbot-1.0.0a8.tar.gz
356,624
87/68/360f279b785ccecbdbfee7f23a815dc3d11ede1cf2e147ae56547036a4c5/ai_reviewbot-1.0.0a8.tar.gz
source
sdist
null
false
408561cf8448d83d9489b85f23396cd1
e25c8fd46829cf8f7867ea9a3b806f1920b7e489e2a93ab530124313684af51f
8768360f279b785ccecbdbfee7f23a815dc3d11ede1cf2e147ae56547036a4c5
null
[ "LICENSE", "NOTICE" ]
200
2.4
archinfo
9.2.202
Classes with architecture-specific information useful to other projects.
# archinfo [![Latest Release](https://img.shields.io/pypi/v/archinfo.svg)](https://pypi.python.org/pypi/archinfo/) [![Python Version](https://img.shields.io/pypi/pyversions/archinfo)](https://pypi.python.org/pypi/archinfo/) [![PyPI Statistics](https://img.shields.io/pypi/dm/archinfo.svg)](https://pypistats.org/packages/archinfo) [![License](https://img.shields.io/github/license/angr/archinfo.svg)](https://github.com/angr/archinfo/blob/master/LICENSE) archinfo is a collection of classes that contain architecture-specific information. It is useful for cross-architecture tools (such as pyvex). ## Project links Project repository: https://github.com/angr/archinfo Documentation: https://api.angr.io/projects/archinfo/en/latest/
text/markdown
null
null
null
null
null
null
[ "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13" ]
[]
null
null
>=3.10
[]
[]
[]
[ "backports-strenum>=1.2.8; python_version < \"3.11\"", "furo; extra == \"docs\"", "myst-parser; extra == \"docs\"", "sphinx; extra == \"docs\"", "sphinx-autodoc-typehints; extra == \"docs\"", "pypcode>=1.1; extra == \"pcode\"", "pytest; extra == \"testing\"", "pytest-xdist; extra == \"testing\"" ]
[]
[]
[]
[ "Homepage, https://api.angr.io/projects/archinfo/en/latest/", "Repository, https://github.com/angr/archinfo" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:44:27.771614
archinfo-9.2.202.tar.gz
40,955
96/2a/6680d2dabfcf6baa276bd7bff79b5513740ff1aad775526ec4f189deb100/archinfo-9.2.202.tar.gz
source
sdist
null
false
45e411b509d1a4cd582bf8a24b8eccd9
edb9a89bfefad5a22260ceb39142e18e91dd4dcf5c1648571535ec0b0f6f84bc
962a6680d2dabfcf6baa276bd7bff79b5513740ff1aad775526ec4f189deb100
BSD-2-Clause
[ "LICENSE" ]
1,216
2.4
angr-management
9.2.202
The official GUI for angr
# angr-management angr-management is a cross-platform, open-source, graphical binary analysis tool powered by the [angr](https://angr.io) binary analysis platform! See [here](https://angr-management.readthedocs.io/en/latest/) for more information. Some screenshots: [![Disassembly](screenshots/disassembly.png)](https://github.com/angr/angr-management/blob/master/screenshots/disassembly.png) [![Decompilation](screenshots/decompilation.png)](https://github.com/angr/angr-management/blob/master/screenshots/decompilation.png)
text/markdown
null
null
null
null
null
null
[ "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only" ]
[]
null
null
>=3.10
[]
[]
[]
[ "PySide6-Essentials!=6.7.0,>=6.4.2", "PySide6-QtAds>=4.2.1", "QtAwesome==1.4.0", "QtPy", "angr[angrDB]==9.2.202", "bidict", "cle==9.2.202", "ipython", "pyqodeng>=0.0.10", "requests[socks]", "tomlkit", "pyobjc-framework-Cocoa; platform_system == \"Darwin\"", "thefuzz[speedup]", "binsync==5.7.11", "rpyc", "qtconsole", "bintrace; extra == \"bintrace\"" ]
[]
[]
[]
[ "Homepage, https://angr.io", "Repository, https://github.com/angr/angr-management" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:44:25.171832
angr_management-9.2.202.tar.gz
3,930,400
60/2c/77aa6b685a523bd19938a326d751463f99036019e649b1a3a5d565cd0a7d/angr_management-9.2.202.tar.gz
source
sdist
null
false
456f58fa0a80021733f5b8356382556e
e4e4765060d720798bfdadcfcb0fc863b3f29ecef4e249502eee25737fbf8deb
602c77aa6b685a523bd19938a326d751463f99036019e649b1a3a5d565cd0a7d
BSD-2-Clause
[ "LICENSE" ]
241
2.4
docfold
0.6.0
Turn any document into structured data. Unified interface for document structuring engines with built-in evaluation.
# Docfold [![PyPI version](https://img.shields.io/pypi/v/docfold.svg)](https://pypi.org/project/docfold/) [![Python 3.10+](https://img.shields.io/badge/python-3.10%2B-blue.svg)](https://www.python.org/downloads/) [![License: MIT](https://img.shields.io/badge/license-MIT-green.svg)](LICENSE) [![CI](https://github.com/mihailorama/docfold/actions/workflows/ci.yml/badge.svg)](https://github.com/mihailorama/docfold/actions/workflows/ci.yml) [![Tests](https://img.shields.io/badge/tests-175%20passed-brightgreen.svg)](#) **Turn any document into structured data.** Unified Python toolkit for document structuring — one interface, 16 engines, built-in benchmarks. > Read the announcement: [Docfold - open-source document processing toolkit](https://datatera.ai/blog/docfold-open-source) ## Engine Comparison > Research-based estimates from public benchmarks, documentation, and community reports. See [detailed methodology](docs/benchmarks.md). Run your own: `docfold compare your_doc.pdf` | Engine | docfold | Type | License | Text PDF | Scan/OCR | Tables | BBox | Conf | Speed | Cost | |--------|:-------:|------|---------|:--------:|:--------:|:------:|:----:|:----:|-------|------| | [**Docling**](https://github.com/docling-project/docling) | ✅ | Local | MIT | ★★★ | ★★☆ | ★★★ | ✅ | — | Medium | Free | | [**MinerU**](https://github.com/opendatalab/MinerU) | ✅ | Local | AGPL | ★★★ | ★★★ | ★★★ | — | — | Slow | Free | | [**Marker**](https://www.datalab.to/) | ✅ | SaaS | Paid | ★★★ | ★★★ | ★★★ | ✅ | — | Fast | $$ | | [**PyMuPDF**](https://pymupdf.readthedocs.io/) | ✅ | Local | AGPL | ★★★ | ☆☆☆ | ★☆☆ | — | — | Ultra | Free | | [**PaddleOCR**](https://github.com/PaddlePaddle/PaddleOCR) | ✅ | Local | Apache | ★☆☆ | ★★★ | ★★☆ | — | ✅ | Medium | Free | | [**Tesseract**](https://github.com/tesseract-ocr/tesseract) | ✅ | Local | Apache | ★☆☆ | ★★☆ | ★☆☆ | — | — | Medium | Free | | [**EasyOCR**](https://github.com/JaidedAI/EasyOCR) | ✅ | Local | Apache | ★☆☆ | ★★★ | ☆☆☆ | — | ✅ | Medium | Free | | [**Unstructured**](https://github.com/Unstructured-IO/unstructured) | ✅ | Local | Apache | ★★☆ | ★★☆ | ★★☆ | — | — | Medium | Free | | [**LlamaParse**](https://docs.llamaindex.ai/en/stable/llama_cloud/llama_parse/) | ✅ | SaaS | Paid | ★★★ | ★★★ | ★★★ | — | — | Fast | $$ | | [**Mistral OCR**](https://docs.mistral.ai/capabilities/document/) | ✅ | SaaS | Paid | ★★★ | ★★★ | ★★★ | — | — | Fast | $$ | | [**Zerox**](https://github.com/getomni-ai/zerox) | ✅ | VLM | MIT | ★★★ | ★★★ | ★★☆ | — | — | Slow | $$$ | | [**AWS Textract**](https://aws.amazon.com/textract/) | ✅ | SaaS | Paid | ★★★ | ★★★ | ★★★ | ✅ | ✅ | Fast | $$ | | [**Google Doc AI**](https://cloud.google.com/document-ai) | ✅ | SaaS | Paid | ★★★ | ★★★ | ★★★ | ✅ | ✅ | Fast | $$ | | [**Azure Doc Intel**](https://azure.microsoft.com/en-us/products/ai-services/ai-document-intelligence) | ✅ | SaaS | Paid | ★★★ | ★★★ | ★★★ | ✅ | ✅ | Fast | $$ | | [**Nougat**](https://github.com/facebookresearch/nougat) | ✅ | Local | MIT | ★★★ | ★★☆ | ★★☆ | — | — | Slow | Free | | [**Surya**](https://github.com/VikParuchuri/surya) | ✅ | Local | GPL | ★★☆ | ★★★ | ★★☆ | ✅ | ✅ | Medium | Free | **★★★** Excellent **★★☆** Good **★☆☆** Basic **☆☆☆** Not supported — **$$** ~$1-3/1K pages **$$$** ~$5-15/1K pages — **BBox** Bounding boxes — **Conf** Confidence scores > [Full engine profiles, format matrix, hardware requirements, and cost breakdown →](docs/benchmarks.md) ## How to Choose | Your situation | Recommended engine | |---|---| | Digital PDF, speed is critical | **PyMuPDF** — zero deps, ~1000 pages/sec | | Scanned documents, need OCR | **PaddleOCR** (80+ langs), **EasyOCR** (PyTorch), or **Tesseract** (100+ langs) | | Complex layouts + tables | **Docling** or **MinerU** (free), **LlamaParse** (paid) | | Academic papers + math formulas | **MinerU** or **Nougat** (free), **Mistral OCR** (paid) | | Best quality, budget available | **Mistral OCR** or **LlamaParse** | | Use any Vision LLM (GPT-4o, Claude, etc.) | **Zerox** — model-agnostic | | Self-hosted, all-in-one ETL | **Unstructured** with hi_res strategy | | Diverse file types (not just PDF) | **Docling** or **Unstructured** | | Need bounding boxes + confidence | **Textract**, **Google DocAI**, or **Azure DocInt** | | Office files (DOCX/PPTX/XLSX) | **Docling**, **Marker**, **Unstructured**, or **Azure DocInt** | | AWS/GCP/Azure native pipeline | **Textract** / **Google DocAI** / **Azure DocInt** | ## Why Docfold? Every engine has trade-offs. Docfold lets you switch between them with one line: | Challenge | Without Docfold | With Docfold | |-----------|----------------|--------------| | Try a new engine | Rewrite your pipeline | Change one string: `engine_hint="docling"` | | Compare quality | Manual side-by-side | `router.compare("doc.pdf")` — one line | | Batch 1000 files | Build your own concurrency | `router.process_batch(files, concurrency=5)` | | Measure accuracy | Write custom metrics | Built-in CER, WER, Table F1, Reading Order | | Switch engines later | Major refactor | Zero code changes — same `EngineResult` | ```python from docfold import EngineRouter from docfold.engines.docling_engine import DoclingEngine from docfold.engines.pymupdf_engine import PyMuPDFEngine router = EngineRouter([DoclingEngine(), PyMuPDFEngine()]) # Auto-select the best available engine result = await router.process("invoice.pdf") print(result.content) # Markdown output print(result.engine_name) # Which engine was used print(result.processing_time_ms) # Compare all engines on the same document results = await router.compare("invoice.pdf") for name, res in results.items(): print(f"{name}: {len(res.content)} chars in {res.processing_time_ms}ms") ``` ## Supported Engines | Engine | Type | License | Formats | GPU | Install | |--------|------|---------|---------|-----|---------| | [**Docling**](https://github.com/docling-project/docling) | Local | MIT | PDF, DOCX, PPTX, XLSX, HTML, images | No | `pip install docfold[docling]` | | [**MinerU**](https://github.com/opendatalab/MinerU) | Local | AGPL-3.0 | PDF | Recommended | `pip install docfold[mineru]` | | [**Marker API**](https://www.datalab.to/) | SaaS | Paid | PDF, Office, images | N/A | `pip install docfold[marker]` | | [**PyMuPDF**](https://pymupdf.readthedocs.io/) | Local | AGPL-3.0 | PDF | No | `pip install docfold[pymupdf]` | | [**PaddleOCR**](https://github.com/PaddlePaddle/PaddleOCR) | Local | Apache-2.0 | Images, scanned PDFs | Optional | `pip install docfold[paddleocr]` | | [**Tesseract**](https://github.com/tesseract-ocr/tesseract) | Local | Apache-2.0 | Images, scanned PDFs | No | `pip install docfold[tesseract]` | | [**EasyOCR**](https://github.com/JaidedAI/EasyOCR) | Local | Apache-2.0 | Images, scanned PDFs | Optional | `pip install docfold[easyocr]` | | [**Unstructured**](https://github.com/Unstructured-IO/unstructured) | Local | Apache-2.0 | PDF, Office, HTML, email, ePub | Optional | `pip install docfold[unstructured]` | | [**LlamaParse**](https://docs.llamaindex.ai/en/stable/llama_cloud/llama_parse/) | SaaS | Paid | PDF, Office, images | N/A | `pip install docfold[llamaparse]` | | [**Mistral OCR**](https://docs.mistral.ai/capabilities/document/) | SaaS | Paid | PDF, images | N/A | `pip install docfold[mistral-ocr]` | | [**Zerox**](https://github.com/getomni-ai/zerox) | VLM | MIT | PDF, images | Depends | `pip install docfold[zerox]` | | [**AWS Textract**](https://aws.amazon.com/textract/) | SaaS | Paid | PDF, images | N/A | `pip install docfold[textract]` | | [**Google Doc AI**](https://cloud.google.com/document-ai) | SaaS | Paid | PDF, images | N/A | `pip install docfold[google-docai]` | | [**Azure Doc Intel**](https://azure.microsoft.com/en-us/products/ai-services/ai-document-intelligence) | SaaS | Paid | PDF, Office, HTML, images | N/A | `pip install docfold[azure-docint]` | | [**Nougat**](https://github.com/facebookresearch/nougat) | Local | MIT (code) | PDF | Recommended | `pip install docfold[nougat]` | | [**Surya**](https://github.com/VikParuchuri/surya) | Local | GPL-3.0 | PDF, images | Optional | `pip install docfold[surya]` | > **Adding your own engine?** Implement the `DocumentEngine` interface — see [Adding a Custom Engine](#adding-a-custom-engine) below. ## Installation ```bash # Core only (no engines — useful for writing custom adapters) pip install docfold # With specific engines pip install docfold[docling] pip install docfold[docling,pymupdf,tesseract] # Everything pip install docfold[all] ``` Requires **Python 3.10+**. ## CLI ```bash # Convert a document docfold convert invoice.pdf docfold convert report.pdf --engine docling --format html --output report.html # List available engines docfold engines # Compare engines on a document docfold compare invoice.pdf # Run evaluation benchmark docfold evaluate tests/evaluation/dataset/ --output report.json ``` ## Batch Processing Process hundreds of documents with bounded concurrency and progress tracking: ```python from docfold import EngineRouter from docfold.engines.docling_engine import DoclingEngine router = EngineRouter([DoclingEngine()]) # Simple batch batch = await router.process_batch( ["invoice1.pdf", "invoice2.pdf", "report.docx"], concurrency=3, ) print(f"{batch.succeeded}/{batch.total} succeeded in {batch.total_time_ms}ms") # With progress callback def on_progress(*, current, total, file_path, engine_name, status, **_): print(f"[{current}/{total}] {status}: {file_path} ({engine_name})") batch = await router.process_batch( file_paths, concurrency=5, on_progress=on_progress, ) # Access results for path, result in batch.results.items(): print(f"{path}: {len(result.content)} chars") # Check errors for path, error in batch.errors.items(): print(f"FAILED {path}: {error}") ``` ## Unified Result Format Every engine returns the same `EngineResult` dataclass: ```python @dataclass class EngineResult: content: str # The extracted text (markdown/html/json/text) format: OutputFormat # markdown | html | json | text engine_name: str # Which engine produced this metadata: dict # Engine-specific metadata pages: int | None # Number of pages processed images: dict | None # Extracted images {filename: base64} tables: list | None # Extracted tables bounding_boxes: list | None # Layout element positions confidence: float | None # Overall confidence [0-1] processing_time_ms: int # Wall-clock time ``` ## Evaluation Framework Docfold includes a built-in evaluation harness to objectively compare engines: ```bash pip install docfold[evaluation] docfold evaluate path/to/dataset/ --engines docling,pymupdf,marker ``` **Metrics measured:** | Metric | What it measures | Target | |--------|------------------|--------| | CER (Character Error Rate) | Character-level text accuracy | < 0.05 | | WER (Word Error Rate) | Word-level text accuracy | < 0.10 | | Table F1 | Table detection and cell content accuracy | > 0.85 | | Heading F1 | Heading detection precision/recall | > 0.90 | | Reading Order Score | Correctness of reading order (Kendall's tau) | > 0.90 | See [docs/evaluation.md](docs/evaluation.md) for the ground truth JSON schema and detailed usage. ## Architecture ``` ┌─────────────────────────────┐ │ Your Application │ └──────────┬──────────────────┘ │ ┌──────────▼──────────────────┐ │ EngineRouter │ │ select() / process() │ │ process_batch() / compare() │ └──────────┬──────────────────┘ │ ┌──────────┬───────┬──────────┴──────┬──────────┬──────────┐ ▼ ▼ ▼ ▼ ▼ ▼ ┌────────┐ ┌────────┐ ┌──────────┐ ┌────────┐ ┌────────┐ ┌──────┐ │Docling │ │ MinerU │ │Unstructd │ │ Marker │ │PyMuPDF │ │ OCR │ │(local) │ │(local) │ │ (local) │ │ (SaaS) │ │(local) │ │Paddle│ └────────┘ └────────┘ └──────────┘ └────────┘ └────────┘ │Tess. │ │ │ │ │ │ └──────┘ │ ┌────────┐ ┌──────────┐ ┌────────┐ │ │ │ │Llama │ │ Mistral │ │ Zerox │ │ │ │ │Parse │ │ OCR │ │ (VLM) │ │ │ │ │(SaaS) │ │ (SaaS) │ │ │ │ │ │ └────────┘ └──────────┘ └────────┘ │ │ │ │ │ │ │ │ │ ┌────────┐ ┌──────────┐ ┌────────┐ │ │ │ │Textract│ │Google │ │ Azure │ │ │ │ │ (AWS) │ │DocAI │ │DocInt │ │ │ │ │ │ │ (GCP) │ │ │ │ │ │ └────────┘ └──────────┘ └────────┘ │ │ └──────────┴───────┴─────────────┴─────────────┴──────────┘ │ ┌────────▼───────┐ │ EngineResult │ │ (unified) │ └────────────────┘ ``` ## Engine Selection Logic When no engine is explicitly specified, the router selects one automatically: 1. **Explicit hint** — `engine_hint="docling"` in the call 2. **Environment default** — `ENGINE_DEFAULT=docling` env var 3. **Extension-aware priority** — each file type has its own engine priority chain (e.g., `.png` prefers PaddleOCR, `.pdf` prefers Docling, `.docx` skips PDF-only engines) 4. **User-configurable** — override with `fallback_order` or restrict with `allowed_engines` ```python # Restrict to specific engines router = EngineRouter(engines, allowed_engines={"docling", "pymupdf"}) # Custom fallback order router = EngineRouter(engines, fallback_order=["pymupdf", "docling", "marker"]) # CLI: --engines flag # docfold convert invoice.pdf --engines docling,pymupdf ``` ## Adding a Custom Engine Implement the `DocumentEngine` interface: ```python from docfold.engines.base import DocumentEngine, EngineResult, OutputFormat class MyEngine(DocumentEngine): @property def name(self) -> str: return "my_engine" @property def supported_extensions(self) -> set[str]: return {"pdf", "docx"} def is_available(self) -> bool: try: import my_library return True except ImportError: return False async def process(self, file_path, output_format=OutputFormat.MARKDOWN, **kwargs): # Your extraction logic here content = extract(file_path) return EngineResult( content=content, format=output_format, engine_name=self.name, ) # Register it router.register(MyEngine()) ``` ## Related Projects Docfold builds on and integrates with these excellent projects: | Project | Description | |---------|-------------| | [Docling](https://github.com/docling-project/docling) | IBM's document conversion toolkit — PDF, DOCX, PPTX, and more | | [MinerU / PDF-Extract-Kit](https://github.com/opendatalab/MinerU) | End-to-end PDF structuring with layout analysis and formula recognition | | [Marker](https://github.com/VikParuchuri/marker) | High-quality PDF to Markdown converter | | [PyMuPDF](https://github.com/pymupdf/PyMuPDF) | Fast PDF/XPS/EPUB processing library | | [PaddleOCR](https://github.com/PaddlePaddle/PaddleOCR) | Multilingual OCR toolkit (80+ languages) | | [Tesseract](https://github.com/tesseract-ocr/tesseract) | Open-source OCR engine (100+ languages) | | [Unstructured](https://github.com/Unstructured-IO/unstructured) | ETL toolkit for diverse document types | | [LlamaParse](https://docs.llamaindex.ai/en/stable/llama_cloud/llama_parse/) | LLM-powered document parsing | | [Mistral OCR](https://docs.mistral.ai/capabilities/document/) | Vision LLM document understanding | | [Zerox](https://github.com/getomni-ai/zerox) | Model-agnostic Vision LLM OCR | | [Nougat](https://github.com/facebookresearch/nougat) | Meta's academic PDF to Markdown model | | [Surya](https://github.com/VikParuchuri/surya) | Multilingual OCR + layout analysis | ### Built by | Project | Description | |---------|-------------| | [Datatera.ai](https://datatera.ai) | AI-powered data transformation and document processing platform | | [Orquesta AI](https://orquestaai.com) | AI orchestration and agent management platform | | [AI Agent Labs](https://aiagentlbs.com) | AI agent services and location-based intelligence | ## Development ```bash git clone https://github.com/mihailorama/docfold.git cd docfold pip install -e ".[dev]" # Run tests pytest # Run linting ruff check src/ tests/ mypy src/ ``` See [CONTRIBUTING.md](CONTRIBUTING.md) for detailed guidelines. ## License MIT. See [LICENSE](LICENSE). > **Note:** Some engine backends have their own licenses (AGPL-3.0 for PyMuPDF and MinerU, GPL-3.0 for Surya, SaaS terms for Marker/LlamaParse/Mistral). Docfold itself is MIT — the engine adapters are optional extras that you install separately.
text/markdown
null
Mihailorama <mihailorama@gmail.com>
null
null
null
document-processing, document-structuring, layout-analysis, ocr, pdf-extraction, table-extraction
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "Intended Audience :: Science/Research", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Text Processing :: General" ]
[]
null
null
>=3.10
[]
[]
[]
[ "azure-ai-documentintelligence>=1.0; extra == \"all\"", "boto3>=1.28; extra == \"all\"", "docling>=2.0; extra == \"all\"", "easyocr>=1.7; extra == \"all\"", "google-cloud-documentai>=3.0; extra == \"all\"", "jiwer>=3.0; extra == \"all\"", "llama-parse>=0.5; extra == \"all\"", "magic-pdf[full]>=0.9; extra == \"all\"", "mistralai>=1.0; extra == \"all\"", "nougat-ocr>=0.1.17; extra == \"all\"", "numpy>=1.23; extra == \"all\"", "opencv-python>=4.7; extra == \"all\"", "paddleocr>=2.7; extra == \"all\"", "paddlepaddle>=2.5; extra == \"all\"", "pdf2image>=1.16; extra == \"all\"", "pillow>=10.0; extra == \"all\"", "psutil>=5.9; extra == \"all\"", "pymupdf>=1.23; extra == \"all\"", "pytesseract>=0.3.10; extra == \"all\"", "requests>=2.31; extra == \"all\"", "scipy>=1.10; extra == \"all\"", "surya-ocr>=0.6; extra == \"all\"", "tabulate>=0.9; extra == \"all\"", "torch>=2.0; extra == \"all\"", "unstructured[all-docs]>=0.16; extra == \"all\"", "azure-ai-documentintelligence>=1.0; extra == \"azure-docint\"", "mypy>=1.8; extra == \"dev\"", "pytest-asyncio>=0.21; extra == \"dev\"", "pytest-cov>=4.0; extra == \"dev\"", "pytest-timeout>=2.1; extra == \"dev\"", "pytest>=7.0; extra == \"dev\"", "ruff>=0.4; extra == \"dev\"", "docling>=2.0; extra == \"docling\"", "easyocr>=1.7; extra == \"easyocr\"", "pdf2image>=1.16; extra == \"easyocr\"", "pillow>=10.0; extra == \"easyocr\"", "jiwer>=3.0; extra == \"evaluation\"", "numpy>=1.23; extra == \"evaluation\"", "psutil>=5.9; extra == \"evaluation\"", "scipy>=1.10; extra == \"evaluation\"", "tabulate>=0.9; extra == \"evaluation\"", "google-cloud-documentai>=3.0; extra == \"google-docai\"", "llama-parse>=0.5; extra == \"llamaparse\"", "requests>=2.31; extra == \"marker\"", "magic-pdf[full]>=0.9; extra == \"mineru\"", "mistralai>=1.0; extra == \"mistral-ocr\"", "nougat-ocr>=0.1.17; extra == \"nougat\"", "torch>=2.0; extra == \"nougat\"", "opencv-python>=4.7; extra == \"paddleocr\"", "paddleocr>=2.7; extra == \"paddleocr\"", "paddlepaddle>=2.5; extra == \"paddleocr\"", "pdf2image>=1.16; extra == \"paddleocr\"", "pillow>=10.0; extra == \"paddleocr\"", "pymupdf>=1.23; extra == \"pymupdf\"", "pillow>=10.0; extra == \"surya\"", "surya-ocr>=0.6; extra == \"surya\"", "torch>=2.0; extra == \"surya\"", "pdf2image>=1.16; extra == \"tesseract\"", "pillow>=10.0; extra == \"tesseract\"", "pytesseract>=0.3.10; extra == \"tesseract\"", "mypy>=1.8; extra == \"test\"", "pytest-asyncio>=0.21; extra == \"test\"", "pytest-cov>=4.0; extra == \"test\"", "pytest-timeout>=2.1; extra == \"test\"", "pytest>=7.0; extra == \"test\"", "ruff>=0.4; extra == \"test\"", "boto3>=1.28; extra == \"textract\"", "unstructured[all-docs]>=0.16; extra == \"unstructured\"", "py-zerox>=0.0.5; extra == \"zerox\"" ]
[]
[]
[]
[ "Homepage, https://github.com/mihailorama/docfold", "Documentation, https://github.com/mihailorama/docfold/tree/main/docs", "Repository, https://github.com/mihailorama/docfold", "Issues, https://github.com/mihailorama/docfold/issues", "Changelog, https://github.com/mihailorama/docfold/blob/main/CHANGELOG.md" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:44:24.789867
docfold-0.6.0.tar.gz
71,777
13/7b/05029e94615127c80d27410b1ed40ed26f6c15bc2058755557bfdf20bd6a/docfold-0.6.0.tar.gz
source
sdist
null
false
8995c512f823ca4315c363d44152a009
06537da13d163db6479a17238c3243644a48f6940597ccbd2d8d0235eb2849cc
137b05029e94615127c80d27410b1ed40ed26f6c15bc2058755557bfdf20bd6a
MIT
[ "LICENSE" ]
208
2.4
angr
9.2.202
A multi-architecture binary analysis toolkit, with the ability to perform dynamic symbolic execution and various static analyses on binaries
# angr [![Latest Release](https://img.shields.io/pypi/v/angr.svg)](https://pypi.python.org/pypi/angr/) [![Python Version](https://img.shields.io/pypi/pyversions/angr)](https://pypi.python.org/pypi/angr/) [![PyPI Statistics](https://img.shields.io/pypi/dm/angr.svg)](https://pypistats.org/packages/angr) [![License](https://img.shields.io/github/license/angr/angr.svg)](https://github.com/angr/angr/blob/master/LICENSE) angr is a platform-agnostic binary analysis framework. It is brought to you by [the Computer Security Lab at UC Santa Barbara](https://seclab.cs.ucsb.edu), [SEFCOM at Arizona State University](https://sefcom.asu.edu), their associated CTF team, [Shellphish](https://shellphish.net), the open source community, and **[@rhelmot](https://github.com/rhelmot)**. ## Project Links Homepage: https://angr.io Project repository: https://github.com/angr/angr Documentation: https://docs.angr.io API Documentation: https://docs.angr.io/en/latest/api.html ## What is angr? angr is a suite of Python 3 libraries that let you load a binary and do a lot of cool things to it: - Disassembly and intermediate-representation lifting - Program instrumentation - Symbolic execution - Control-flow analysis - Data-dependency analysis - Value-set analysis (VSA) - Decompilation The most common angr operation is loading a binary: `p = angr.Project('/bin/bash')` If you do this in an enhanced REPL like IPython, you can use tab-autocomplete to browse the [top-level-accessible methods](https://docs.angr.io/core-concepts/toplevel) and their docstrings. The short version of "how to install angr" is `mkvirtualenv --python=$(which python3) angr && python -m pip install angr`. ## Example angr does a lot of binary analysis stuff. To get you started, here's a simple example of using symbolic execution to get a flag in a CTF challenge. ```python import angr project = angr.Project("angr-doc/examples/defcamp_r100/r100", auto_load_libs=False) @project.hook(0x400844) def print_flag(state): print("FLAG SHOULD BE:", state.posix.dumps(0)) project.terminate_execution() project.execute() ``` # Quick Start - [Install Instructions](https://docs.angr.io/introductory-errata/install) - Documentation as [HTML](https://docs.angr.io/) and sources in the angr [Github repository](https://github.com/angr/angr/tree/master/docs) - Dive right in: [top-level-accessible methods](https://docs.angr.io/core-concepts/toplevel) - [Examples using angr to solve CTF challenges](https://docs.angr.io/examples). - [API Reference](https://docs.angr.io/en/latest/api.html) - [awesome-angr repo](https://github.com/degrigis/awesome-angr)
text/markdown
null
null
null
null
null
null
[ "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13" ]
[]
null
null
>=3.10
[]
[]
[]
[ "cxxheaderparser", "GitPython", "archinfo==9.2.202", "cachetools", "capstone==5.0.6", "cffi>=1.14.0", "claripy==9.2.202", "cle==9.2.202", "lmdb", "msgspec", "mulpyplexer", "networkx!=2.8.1,>=2.0", "protobuf>=6.33.0", "psutil", "pycparser~=3.0", "pydemumble", "pypcode<4.0,>=3.2.1", "pyvex==9.2.202", "rich>=13.1.0", "sortedcontainers", "sympy", "typing-extensions", "colorama; platform_system == \"Windows\"", "sqlalchemy; extra == \"angrdb\"", "keystone-engine; extra == \"keystone\"", "opentelemetry-api; extra == \"telemetry\"", "litellm; extra == \"llm\"", "unicorn==2.1.4; extra == \"unicorn\"" ]
[]
[]
[]
[ "Homepage, https://angr.io/", "Repository, https://github.com/angr/angr" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:44:23.483887
angr-9.2.202.tar.gz
4,094,813
61/e6/6dbbed86856035bc742b9d405c0aab0e42248ddb2b7ca07a85bb7a93173d/angr-9.2.202.tar.gz
source
sdist
null
false
7cbcd020270f38aa66a3c65c206bd3fb
0cb92bb222aa19ab3e008c3043b1d7b224b50d40b6a987c21d30ac03f94b6eff
61e66dbbed86856035bc742b9d405c0aab0e42248ddb2b7ca07a85bb7a93173d
BSD-2-Clause
[ "LICENSE" ]
3,063
2.4
cerc-persistence
2.0.0.2
CERC Persistence consist of a set of classes to store and retrieve Cerc Hub cities and results
CERC Persistence consist of a set of classes to store and retrieve Cerc Hub cities and results. Developed at Concordia university in Canada as part of the research group from the Next Generation Cities Institute, our aim among others is to provide a comprehensive set of tools to help researchers and urban developers to make decisions to improve the livability and efficiency of our cities, cerc persistence will store the simulation results and city information to make those available for other researchers.
null
null
null
null
null
null
null
[ "License :: OSI Approved :: GNU Library or Lesser General Public License (LGPL)", "Programming Language :: Python", "Programming Language :: Python :: 3" ]
[]
null
null
null
[]
[]
[]
[ "python-dotenv", "sqlalchemy", "cerc-hub", "psycopg2-binary", "geoalchemy2", "setuptools" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.12.9
2026-02-20T21:44:00.481252
cerc_persistence-2.0.0.2.tar.gz
16,954
dd/10/119c6f004cc4b559b5d6e045a044319046e019bce18d43a118628048bdae/cerc_persistence-2.0.0.2.tar.gz
source
sdist
null
false
cb18502a0038da161387fae721dce232
97e885bb124ce7e1fe665bf388b9ae8b5756552cb0bc057c1839d4e6062ddac6
dd10119c6f004cc4b559b5d6e045a044319046e019bce18d43a118628048bdae
null
[ "LICENSE.md" ]
219
2.4
quadrants
0.4.1
The Quadrants Programming Language
# What is Quadrants? Quadrants is a high-performance multi-platform compiler for physics simulation being continuously developed by [Genesis AI](https://genesis-ai.company/). It is designed for large-scale physics simulation and robotics workloads. It compiles Python code into highly optimized parallel kernels that run on: * NVIDIA GPUs (CUDA) * Vulkan-compatible GPUs (SPIR-V) * Apple Metal GPUs * AMD GPUs (ROCm HIP) * x86 and ARM64 CPUs ## The origin The quadrants project was originally forked from [Taichi](https://github.com/taichi-dev/taichi) in June 2025. As the original Taichi is no longer being maintained and the codebase evolved into a fully independent compiler with its own direction and long-term roadmap, we decided to give it a name that reflects both its roots and its new identity. The name _Quadrants_ is inspired by the Chinese saying: > 太极生两仪,两仪生四象 > > The Supreme Polarity (Taichi) gives rise to the Two Modes (Ying & Yang), which in turn give rise to the Four Forms (_Quadrants_). _Quadrants_ captures the idea of progression originated from taichi — built on the same foundation, evolving in its own direction while acknowledging its roots. This project is now fully independent and does not aim to maintain backward compatibility with upstream Taichi. ## How Quadrants differs from upstream Taichi While the repository still resembles upstream in structure, major changes include: ### Modernized infrastructure * Revamped CI * Support for Python 3.10–3.13 * Support for macOS up to 15 * Significantly improved reliability (≥90% CI success on correct code) ### Structural improvements * Added `dataclasses.dataclass` structs: * Work with both ndarrays and fields * Can be passed into child `ti.func` functions * Can be nested * No kernel runtime overhead (kernels see only underlying arrays) ### Removed components To focus the compiler and reduce maintenance burden, we removed: * GUI / GGUI * C-API * AOT * DX11 / DX12 * iOS / Android * OpenGL / GLES * argpack * CLI ### Performance improvements #### Reduced launch latency * Release 4.0.0 improved non-batched ndarray CPU performance by **4.5×** in Genesis benchmarks. * Release 3.2.0 improved ndarray performance from **11× slower than fields** to **1.8× slower** (on a 5090 GPU, Genesis benchmark). #### Reduced warm-cache latency On Genesis simulator (Linux + NVIDIA 5090): * `single_franka_envs.py` cache load time reduced from **7.2s → 0.3s** #### Zero-copy Torch interop * Added `to_dlpack` * Enables zero-copy memory sharing between PyTorch and Quadrants * Avoids kernel-based accessors * Significantly improves performance ### Compiler upgrades * Upgraded to LLVM 20 * Enabled ARM support --- # Installation ## Prerequisites - Python 3.10-3.13 - Mac OS 14, 15, Windows, or Ubuntu 22.04-24.04 or compatible ## Procedure ``` pip install quadrants ``` (For how to build from source, see our CI build scripts, e.g. [linux build scripts](.github/workflows/scripts_new/linux_x86/) ) # Documentation - [docs](https://genesis-embodied-ai.github.io/quadrants/user_guide/index.html) - [API reference](https://genesis-embodied-ai.github.io/quadrants/autoapi/index.html) # Something is broken! - [Create an issue](https://github.com/Genesis-Embodied-AI/quadrants/issues/new/choose) # Acknowledgements Quadrants stands on the shoulders of the original [Taichi](https://github.com/taichi-dev/taichi) project, built with care and vision by many contributors over the years. For the full list of contributors and credits, see the [original Taichi repository](https://github.com/taichi-dev/taichi). We are grateful for that foundation.
text/markdown
null
null
Quadrants developers
null
Apache-2.0
graphics, simulation
[ "Development Status :: 5 - Production/Stable", "Topic :: Software Development :: Compilers", "Topic :: Multimedia :: Graphics", "Topic :: Games/Entertainment :: Simulation", "Intended Audience :: Science/Research", "Intended Audience :: Developers", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13" ]
[]
null
null
<3.14,>=3.10
[]
[]
[]
[ "numpy>=1.26.0", "colorama", "dill", "pydantic>=2.0.0", "rich>=1.0.0", "setuptools>=77.0.0", "cffi>=1.16.0" ]
[]
[]
[]
[ "Homepage, https://github.com/Genesis-Embodied-AI/quadrants" ]
twine/6.2.0 CPython/3.10.19
2026-02-20T21:43:35.303510
quadrants-0.4.1-cp313-cp313-macosx_11_0_arm64.whl
30,452,309
2f/37/6ae8b0e5b29b2d199fd569dc4f0970f006dbd63bd3ed7345d22321d2640f/quadrants-0.4.1-cp313-cp313-macosx_11_0_arm64.whl
cp313
bdist_wheel
null
false
8a4af11bb3bc8ab6919beba8b2143960
ba68e6203e6ae87e0637170ff9ad6c9f48d339c426f6932d345231417bff2762
2f376ae8b0e5b29b2d199fd569dc4f0970f006dbd63bd3ed7345d22321d2640f
null
[ "LICENSE" ]
899
2.4
e2llm-medsynth
0.1.0
Synthetic medical record generator with realistic schema variance across locales
# MedSynth [![PyPI](https://img.shields.io/pypi/v/e2llm-medsynth)](https://pypi.org/project/e2llm-medsynth/) [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/licenses/MIT) [![Python 3.11+](https://img.shields.io/badge/python-3.11+-blue.svg)](https://www.python.org/downloads/) [![Tests](https://github.com/e2llm/medsynth/actions/workflows/test.yml/badge.svg)](https://github.com/e2llm/medsynth/actions/workflows/test.yml) Multi-lingual synthetic healthcare data generator. Produces realistic medical records with intentional OCR artifacts and schema variance — simulating real-world messy healthcare data. ## The Problem Healthcare AI development is bottlenecked by data access. Real patient records are legally restricted (HIPAA, GDPR, Uruguay's Ley 18.331), expensive to anonymize, and nearly impossible to share across borders. Researchers spend months navigating data access before writing a single line of AI code. Meanwhile, most synthetic data generators produce clean, English-only records that look nothing like actual hospital data — which is scanned paper, multi-lingual, inconsistently formatted, and full of OCR errors. MedSynth generates data that looks like the real thing — including the mess. ## What Makes This Different | Feature | MedSynth | Typical Generators | |---------|----------|-------------------| | **Languages** | 6 locales (Hebrew, Arabic, Spanish) | English only | | **OCR artifacts** | Realistic scan errors per script | Clean text | | **Schema variance** | Different formats per facility | Single schema | | **ID systems** | Country-specific (Teudat Zehut, CURP, DNI) | Generic | | **Privacy** | Zero real patient data | Often derived from real records | ### OCR Realism Real medical records are scanned paper. MedSynth simulates actual scanning artifacts: - **Arabic**: Dot-group confusions (ب↔ت↔ث), tashkeel stripping - **Hebrew**: Shape-based confusions (ר↔ד, ח↔כ) - **Latin**: rn→m merges, diacritic loss (ñ→n), 0↔O swaps ### Schema Variance Different hospitals format records differently. MedSynth produces variant schemas across facilities so AI systems learn to handle real-world inconsistency — not just clean demos. ## Installation ```bash pip install e2llm-medsynth ``` ### Quick Start ```bash pip install e2llm-medsynth # Structured data only (no LLM needed) medsynth --locale he_IL --num-patients 10 --skip-freetext -v # With free text via Ollama (default — no API key needed) ollama pull llama4:maverick medsynth --locale he_IL --num-patients 10 -v ``` Free text generation uses any OpenAI-compatible API. Default: Ollama + Llama 4 Maverick (local). No API key needed for basic generation or local Ollama. ## Output Format MedSynth outputs NDJSON files — one per facility × document type: ``` output/ ├── medical_alon_discharge.ndjson ├── medical_alon_lab.ndjson ├── medical_alon_referral.ndjson ├── medical_hadarim_discharge.ndjson ├── medical_hadarim_visit.ndjson ├── ... ``` ### Example: The Mess **Alon hospital — digital, English field names:** ```json {"patient_id": "165667015", "patient_name": "משה אזולאי", "patient_age": 77, "gender": "male", "document_date": "2023-07-01", "facility_name": "בית חולים האלון", "conditions": ["השמנת יתר", "דיכאון", "COPD"], "smoking_status": true, "department": "אורולוגיה", "primary_diagnosis": "השמנת יתר", "doc_type": "discharge"} ``` **Hadarim hospital — OCR source, Hebrew field names, different ID type:** ```json {"מספר_זהות": 161559406, "שם_מטופל": "יעל גולן", "גיל": 31, "מין": "female", "תאריך": "29/04/2024", "מוסד_רפואי": "מרכז רפואי הדרים", "מחלות_רקע": ["סוכרת סוג 2", "אי ספיקת כליות כרונית"], "מחלקה": "פנימית א", "אבחנה_ראשית": "סוכרת סוג 2", "doc_type": "discharge"} ``` Different field names (`patient_id` → `מספר_זהות`), different date format (`2023-07-01` → `29/04/2024`), ID as integer instead of string. **Saudi Arabia — Arabic fields, age as range string:** ```json {"رقم_الهوية": 1496965326, "الاسم": "عبدالرحمن بن راشد الأحمدي", "العمر": "50-60", "الجنس": "male", "التاريخ": "2023-06", "المركز": "مركز الرعاية الصحية الأولية", "الأمراض": ["فرط شحميات الدم"], "التشخيص": "فرط شحميات الدم", "doc_type": "discharge"} ``` Age stored as range string (`"50-60"` not `57`), date truncated to month (`"2023-06"`). **Mexico — CURP national ID, Spanish field names:** ```json {"patient_id": "AULJ460528MDFGPN03", "patient_name": "Juana Aguilar Figueroa", "patient_age": 77, "gender": "female", "document_date": "2024-01-10", "facility_name": "Hospital Nacional del Norte", "conditions": ["insuficiencia renal crónica", "obesidad", "gota"], "department": "oncología", "doc_type": "discharge"} ``` 18-character CURP encodes name, DOB, gender, and state — completely different from Israeli 9-digit Luhn IDs. ## CLI Usage ```bash # Default: Ollama + Llama 4 Maverick (local, no API key) medsynth --locale he_IL --num-patients 500 --seed 42 -v # Structured data only — no LLM needed medsynth --locale es_MX --num-patients 50 --seed 42 --skip-freetext -v # OpenAI GPT-4o export LLM_API_KEY="sk-..." medsynth --api-base https://api.openai.com/v1 --model gpt-4o -v # Moonshot Kimi K2 export LLM_API_KEY="your-moonshot-key" medsynth --api-base https://api.moonshot.ai/v1 --model kimi-k2-0711-preview -v # Anthropic Claude Haiku (via LiteLLM or any OpenAI-compatible proxy) medsynth --api-base http://localhost:4000/v1 --model claude-haiku-4-5 -v ``` ### Options | Flag | Default | Description | |------|---------|-------------| | `--locale` | `he_IL` | Locale code | | `--num-patients` | `500` | Number of patients to generate | | `--seed` | `42` | Random seed for reproducibility | | `--output-dir` | `output` | Output directory for NDJSON files | | `--model` | `llama4:maverick` | LLM model name | | `--api-base` | `http://localhost:11434/v1` | API base URL (any OpenAI-compatible endpoint) | | `--api-key` | — | API key (or set `LLM_API_KEY` / `OPENAI_API_KEY` env var) | | `--skip-freetext` | off | Skip LLM calls for free text | | `--force` | off | Overwrite existing output files | | `-v` / `--verbose` | off | Verbose output | ## Python API ```python from medsynth import generate_documents, load_locale # Generate documents (default: Ollama + Llama 4 Maverick) counts = generate_documents( num_patients=50, seed=42, output_dir="output", locale_code="es_ES", skip_freetext=True, # set False to generate free text via LLM verbose=True, ) # Use a different provider counts = generate_documents( num_patients=50, seed=42, output_dir="output", model="gpt-4o", api_base="https://api.openai.com/v1", api_key="sk-...", locale_code="es_ES", ) # Load a locale directly locale = load_locale("ar_SA") print(locale.code, len(locale.facilities)) ``` ## Supported Locales | Code | Region | Script | Facilities | |------|--------|--------|------------| | `he_IL` | Israel | Hebrew | Alon, Hadarim, Shaked, Ofek | | `ar_SA` | Saudi Arabia | Arabic | Riyadh Medical City, Royal Military, PHC, Al Hayat Labs | | `ar_EG` | Egypt | Arabic | Nile Central, Delta University, Tahrir, Al Mokhtabar | | `es_ES` | Spain | Latin | Reina Ficticia, San Rafael, Atencion Primaria, Iberia Labs | | `es_MX` | Mexico | Latin | Nacional del Norte, Federal del Centro, Centro de Salud, Azteca Labs | | `es_AR` | Argentina | Latin | Hospital del Plata, San Martin, CAPS, Austral Labs | ## Sample Data Pre-generated sample data (50 patients, seed 42) ships with the package: ```python from importlib.resources import files sample_dir = files("medsynth") / "sample_data" / "he_IL" ``` ## Tests ```bash pip install -e ".[dev]" pytest tests/ -v ``` ## Use Cases - **Healthcare NLP testing** — validate extraction pipelines against known-correct synthetic records - **AI agent development** — train/test agents that query unstructured medical text - **OCR pipeline validation** — test document understanding against realistic scan artifacts - **Cross-border healthcare IT** — test systems handling multiple languages/formats - **Compliance testing** — validate anonymization systems with synthetic ground truth - **Education** — teach healthcare informatics without privacy concerns ## Who We Are [e2llm](https://github.com/e2llm) — healthcare data intelligence. We build systems that make unstructured medical data queryable: document understanding (OCR → structured), semantic search (natural language → patient cohorts), and multi-lingual medical NLP. Working with healthcare organizations across MENA and Latin America. ## Contact - **Email**: info@e2llm.com - **For**: Custom locale development, integration with production pipelines, air-gapped deployment consulting, enterprise support ## Contributing PRs welcome. See [issues](https://github.com/e2llm/medsynth/issues) for open tasks. ## Disclaimer **MedSynth is an independent open-source project by [e2llm](https://github.com/e2llm). It is not affiliated with, endorsed by, or related to any company or entity operating under the same or a similar name.** Any resemblance in naming is purely coincidental. This tool generates entirely synthetic data for software testing, demos, and research. No real patient information is used or produced. Facility names are fictional — inspired by real institutions for realism, but all generated records are entirely synthetic. This is not medical software and must not be used for clinical decisions. Free text generation calls an LLM API. The default (Ollama) runs locally at no cost. When using cloud providers (OpenAI, Moonshot, Anthropic), review their usage policies and be aware of associated costs. ## License MIT
text/markdown
null
null
null
null
null
synthetic-data, medical-records, ehr, elasticsearch, testing
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "Intended Audience :: Science/Research", "Programming Language :: Python :: 3", "Topic :: Scientific/Engineering :: Medical Science Apps.", "Topic :: Software Development :: Testing" ]
[]
null
null
>=3.11
[]
[]
[]
[ "openai>=1.0", "python-dotenv", "pytest>=7.0; extra == \"dev\"" ]
[]
[]
[]
[ "Repository, https://github.com/e2llm/medsynth" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:43:25.234374
e2llm_medsynth-0.1.0.tar.gz
1,106,827
23/ca/955150820782df7441c183dc9a88ede7e7f3f93057f5c78e2309d8af79bf/e2llm_medsynth-0.1.0.tar.gz
source
sdist
null
false
787893c91f5eb38b8c53884cb8626bf4
29daf830e69fb8e8bd6625bd7dea94045b5eeecd9c86baeb19ee8ab56ab9e0b5
23ca955150820782df7441c183dc9a88ede7e7f3f93057f5c78e2309d8af79bf
MIT
[ "LICENSE" ]
231
2.4
senpuki
0.3.0
Distributed Durable Functions in Python
# Senpuki Distributed durable functions for Python. Write reliable, stateful workflows using async/await. ```bash pip install senpuki ``` ## Quick Example ```python import asyncio from senpuki import Senpuki, Result @Senpuki.durable() async def process_order(order_id: str) -> dict: await asyncio.sleep(1) # Simulate work return {"order_id": order_id, "status": "processed"} @Senpuki.durable() async def order_workflow(order_ids: list[str]) -> Result[list, Exception]: results = [] for order_id in order_ids: result = await process_order(order_id) results.append(result) return Result.Ok(results) async def main(): backend = Senpuki.backends.SQLiteBackend("workflow.db") await backend.init_db() executor = Senpuki(backend=backend) worker = asyncio.create_task(executor.serve()) exec_id = await executor.dispatch(order_workflow, ["ORD-001", "ORD-002"]) result = await executor.wait_for(exec_id) print(result.value) asyncio.run(main()) ``` ## Why Senpuki? | Feature | Temporal | Celery | Prefect | Airflow | **Senpuki** | |---------|----------|--------|---------|---------|-------------| | Durable Execution | Yes | No | Partial | No | **Yes** | | Setup Complexity | High | Medium | Medium | High | **Very Low** | | Infrastructure | Server cluster | Broker | Server | Multi-component | **SQLite/Postgres** | | Native Async | Yes | No | Yes | Limited | **Yes** | Senpuki fills the gap between simple task queues (Celery) and enterprise platforms (Temporal): - **vs Temporal**: Same durability guarantees, fraction of the infrastructure - **vs Celery/Dramatiq**: True workflow durability, not just task retries - **vs Prefect/Airflow**: Application workflows, not batch data pipelines See [full comparison](docs/comparison.md) for details. ## Features - **Durable Execution** - Workflow state survives crashes and restarts - **Automatic Retries** - Configurable retry policies with exponential backoff - **Distributed Workers** - Scale horizontally across multiple processes - **Parallel Execution** - Fan-out/fan-in with `asyncio.gather` and `Senpuki.map` - **Rate Limiting** - Control concurrent executions per function - **External Signals** - Coordinate workflows with external events - **Dead Letter Queue** - Inspect and replay failed tasks - **Idempotency & Caching** - Prevent duplicate work - **Multiple Backends** - SQLite (dev) or PostgreSQL (production) - **OpenTelemetry** - Distributed tracing support ## Key Concepts ```python from senpuki import Senpuki, RetryPolicy # Configurable retry policies @Senpuki.durable( retry_policy=RetryPolicy(max_attempts=5, initial_delay=1.0), queue="high_priority", max_concurrent=10, # Rate limiting idempotent=True, # Prevent duplicate execution ) async def my_activity(data: dict) -> dict: ... # Durable sleep (doesn't block workers) await Senpuki.sleep("30m") # Parallel execution results = await asyncio.gather(*[process(item) for item in items]) # Or optimized for large batches: results = await Senpuki.map(process, items) # External signals payload = await Senpuki.wait_for_signal("approval") await executor.send_signal(exec_id, "approval", {"approved": True}) ``` ## Backends ```python # SQLite (development) backend = Senpuki.backends.SQLiteBackend("senpuki.db") # PostgreSQL (production) backend = Senpuki.backends.PostgresBackend("postgresql://user:pass@host/db") # Optional: Redis for low-latency notifications executor = Senpuki( backend=backend, notification_backend=Senpuki.notifications.RedisBackend("redis://localhost") ) ``` ## CLI ```bash senpuki list # List executions senpuki show <exec_id> # Show execution details senpuki dlq list # List dead-lettered tasks senpuki dlq replay <task_id> # Replay failed task ``` ## Documentation Full documentation available in [`docs/`](docs/): - [Getting Started](docs/getting-started.md) | [Core Concepts](docs/core-concepts.md) | [Comparison](docs/comparison.md) - **Guides**: [Durable Functions](docs/guides/durable-functions.md) | [Orchestration](docs/guides/orchestration.md) | [Error Handling](docs/guides/error-handling.md) | [Parallel Execution](docs/guides/parallel-execution.md) | [Signals](docs/guides/signals.md) | [Workers](docs/guides/workers.md) | [Monitoring](docs/guides/monitoring.md) - **Patterns**: [Saga](docs/patterns/saga.md) | [Batch Processing](docs/patterns/batch-processing.md) - **Reference**: [API](docs/api-reference/senpuki.md) | [Configuration](docs/configuration.md) | [Deployment](docs/deployment.md) ## Examples See [`examples/`](examples/) for complete workflows: - `simple_flow.py` - Basic workflow - `saga_trip_booking.py` - Saga pattern with compensation - `batch_processing.py` - Fan-out/fan-in - `media_pipeline.py` - Complex multi-stage pipeline ## Requirements - Python 3.12+ - `aiosqlite` or `asyncpg` (backend) - `redis` (optional, for notifications) ## License MIT
text/markdown
null
noku <noku@onlypa.ws>
null
null
null
null
[]
[]
null
null
>=3.12
[]
[]
[]
[ "aiosqlite>=0.21.0", "asyncpg>=0.31.0", "redis>=7.1.0" ]
[]
[]
[]
[ "Homepage, https://github.com/nokusukun/senpuki", "Issues, https://github.com/nokusukun/senpuki/issues" ]
twine/6.2.0 CPython/3.13.5
2026-02-20T21:43:19.662968
senpuki-0.3.0.tar.gz
78,979
51/2b/678e3309c79777ca09ea9899bf1d40ebec7de51a3eab641149cfa8234e54/senpuki-0.3.0.tar.gz
source
sdist
null
false
b7be081697c8293490ba5c1a9824e5b6
462a68074add7b066906af5fba030362015e10323b5f75bfdac801defea1ddf4
512b678e3309c79777ca09ea9899bf1d40ebec7de51a3eab641149cfa8234e54
null
[]
229
2.4
iatoolkit
1.56.1
IAToolkit
# 🧠 IAToolkit — Open-Source Framework for Real-World AI Assistants Build private, production-grade AI assistants that run entirely inside your environment and speak the language of your business. IAToolkit is not a demo wrapper or a prompt playground — it is a **full architecture** for implementing intelligent systems that combine LLMs, SQL data, internal documents, tools, workflows, and multi-tenant business logic. --- ## 🚀 Why IAToolkit? Modern AI development is fragmented: LangChain handles chains, LlamaIndex handles documents, your backend handles SQL, your frontend handles chats, and your devs glue everything together. **IAToolkit brings all of this into one unified, production-ready framework.** It focuses on: - **real-world data** (SQL + documents) - **real workflows** (LLM tools + python services) - **real multi-tenant architecture** (1 company → many companies) - **real constraints** (security, reproducibility, governance) - **real deployment** (your servers, your infrastructure) IAToolkit lets you build the assistant that *your* organization needs — not a generic chatbot. --- ## 🧩 Architecture in a Nutshell IAToolkit is a structured, layered framework: Interfaces (Web & API) ↓ Intelligence Layer (prompts, tools, SQL orchestration, RAG) ↓ Execution Layer (services, workflows, validation) ↓ Data Access (SQLAlchemy, connectors) ↓ Company Modules (company.yaml + custom tools) ### ✔ Interfaces Chat UI, REST API, auth, sessions, JSON/HTML responses. ### ✔ Intelligence Layer Core logic: prompt rendering, SQL orchestration, RAG, LLM tool dispatching. ### ✔ Execution Layer Python services that implement real workflows: querying data, generating reports, retrieving documents, executing business logic. ### ✔ Data Access A clean repository pattern using SQLAlchemy. ### ✔ Company Modules Each company has: - its own `company.yaml` - its own prompts - its own tools - its own services - its own vector store & SQL context This modularity allows **true multi-tenancy**. --- ## 🔌 Connect to Anything IAToolkit integrates naturally with: - **SQL databases** (PostgreSQL, MySQL, SQL Server, etc.) - **Document retrieval** (PDF, text, embeddings) - **External APIs** - **Internal microservices** - **Custom Python tools** It also includes a **production-grade RAG pipeline**, combining: - embeddings - chunking - hybrid search - SQL queries + document retrieval - tool execution Everything orchestrated through the Intelligence Layer. --- ## 🏢 Multi-Tenant Architecture A single installation of IAToolkit can power assistants for multiple companies, departments, or customers. ```text companies/ company_a company_b company_c ``` Each Company is fully isolated: - prompts - tools - credentials - documents - SQL contexts - business rules This makes IAToolkit ideal for SaaS products, agencies, consultancies, and organizations with multiple business units. --- ## 🆓 Community Edition vs Enterprise Edition IAToolkit follows a modern **open-core** model: ### 🟦 Community Edition (MIT License) - Full Open-Source Core - SQL + Basic RAG - One Company - Custom Python tools - Self-managed deployment Perfect for developers, small teams, single-business use cases, and experimentation. ### 🟥 Enterprise Edition (Commercial License) - Unlimited Companies (multi-tenant) - Payment services integration - Enterprise Agent Workflows - SSO integration - Priority support & continuous updates - Activation via **License Key** 👉 Licensing information: - [Community Edition (MIT)](LICENSE_COMMUNITY.md) - [Enterprise License](ENTERPRISE_LICENSE.md) --- ## 🧩 Who Is IAToolkit For? - Companies building internal “ChatGPT for the business” - SaaS products adding AI assistants for multiple customers - AI teams that need reproducible prompts and controlled tools - Developers who want real workflows, not toy demos - Organizations requiring privacy, security, and self-hosting - Teams working with SQL-heavy business logic - Consultancies deploying AI for multiple clients --- ## ⭐ Key Differentiators - prioritizes **architecture-first design**, not chains or wrappers - supports **multi-company** out of the box - integrates **SQL, RAG, and tools** into a single intelligence layer - keeps **business logic isolated** inside Company modules - runs entirely **on your own infrastructure** - ships with a **full web chat**, and API. - is built for **production**, not prototypes --- ## 📚 Documentation - 🚀 **[Quickstart](docs/quickstart.md)** – Set up your environment and run the project - ☁️ **[Deployment Guide](docs/deployment_guide.md)** – Production deployment instructions - 🏗️ **[Companies & Components](docs/companies_and_components.md)** – how Company modules work - 🧠 **[Programming Guide](docs/programming_guide.md)** – services, intelligence layer, dispatching - 🗃️ **[Database Guide](docs/database_guide.md)** – internal schema overview - 🌱 **[Foundation Article](https://iatoolkit.com/pages/foundation)** – the “Why” behind the architecture - 📘 **[Mini-Project (3 months)](https://iatoolkit.com/pages/mini_project)** – how to deploy a corporate AI assistant --- ## 🤝 Contributing IAToolkit is open-source and community-friendly. PRs, issues, ideas, and feedback are always welcome. --- ## ⭐ Support the Project If you find IAToolkit useful, please **star the GitHub repo** — it helps visibility and adoption.
text/markdown
Fernando Libedinsky
null
null
null
null
null
[]
[]
null
null
>=3.12
[]
[]
[]
[ "bcrypt==4.2.1", "boto3==1.36.22", "botocore==1.36.22", "build==1.2.2.post1", "click==8.1.8", "cryptography==44.0.3", "Flask==3.1.0", "Flask-Bcrypt==1.0.1", "flask-cors==6.0.0", "Flask-Injector==0.15.0", "Flask-Session==0.8.0", "flatbuffers==24.3.25", "google-ai-generativelanguage==0.6.15", "google-api-core==2.24.1", "google-api-python-client==2.161.0", "google-auth>=2.46.0", "google-auth-httplib2==0.2.0", "google-auth-oauthlib==1.2.1", "google-cloud-core==2.4.1", "google-cloud-storage==3.0.0", "google-crc32c==1.6.0", "google-genai==1.57.0", "google-resumable-media==2.7.2", "googleapis-common-protos==1.66.0", "gunicorn==23.0.0", "h11==0.14.0", "httpcore==1.0.7", "httplib2==0.22.0", "httptools==0.6.4", "httpx<1.0.0,>=0.28.1", "httpx-sse==0.4.0", "huggingface-hub>=0.16.4", "humanfriendly==10.0", "idna==3.10", "injector==0.22.0", "Jinja2==3.1.5", "langchain==0.3.19", "langchain-core==0.3.35", "langchain-text-splitters==0.3.6", "markdown2==2.5.3", "openai==2.8.1", "openpyxl==3.1.5", "pandas==2.3.1", "pgvector==0.3.6", "pillow==11.0.0", "psutil==7.0.0", "psycopg2-binary==2.9.10", "PyJWT==2.10.1", "PyMuPDF==1.25.0", "python-dotenv==1.0.1", "pytest==8.3.4", "pytest-cov==5.0.0", "pytest-mock==3.14.0", "python-dateutil==2.9.0.post0", "python-docx==1.1.2", "pytesseract==0.3.13", "pytz==2025.2", "PyYAML==6.0.2", "redis==5.2.1", "regex==2024.11.6", "requests==2.32.3", "requests-oauthlib==2.0.0", "requests-toolbelt==1.0.0", "s3transfer==0.11.2", "sib-api-v3-sdk==7.6.0", "SQLAlchemy==2.0.36", "tiktoken==0.8.0", "tokenizers>=0.22.0", "websocket-client==1.8.0", "websockets==14.1", "Werkzeug==3.1.3", "pyjwt[crypto]>=2.8.0", "docling==2.72.0; extra == \"docling\"" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:43:03.892200
iatoolkit-1.56.1.tar.gz
440,207
a4/8a/5f8055ded8cdf2283edf8dd93e3f359d5a62b2c035307cc14866e5ee2aa3/iatoolkit-1.56.1.tar.gz
source
sdist
null
false
4a13517243294d3783c764d1c0f81851
f1d8ce60adb034a70129456d7504f2242f74c6a1a8a6daa8dd8fa40df737f651
a48a5f8055ded8cdf2283edf8dd93e3f359d5a62b2c035307cc14866e5ee2aa3
MIT
[ "LICENSE", "LICENSE_COMMUNITY.md" ]
206
2.4
pico-fastapi
0.3.0
Pico-ioc integration for FastAPI. Adds Spring Boot-style controllers, autoconfiguration, and scopes (request, websocket, session).
# 📦 pico-fastapi [![PyPI](https://img.shields.io/pypi/v/pico-fastapi.svg)](https://pypi.org/project/pico-fastapi/) [![Ask DeepWiki](https://deepwiki.com/badge.svg)](https://deepwiki.com/dperezcabrera/pico-fastapi) [![License: MIT](https://img.shields.io/badge/License-MIT-blue.svg)](https://opensource.org/licenses/MIT) ![CI (tox matrix)](https://github.com/dperezcabrera/pico-fastapi/actions/workflows/ci.yml/badge.svg) [![codecov](https://codecov.io/gh/dperezcabrera/pico-fastapi/branch/main/graph/badge.svg)](https://codecov.io/gh/dperezcabrera/pico-fastapi) [![Quality Gate Status](https://sonarcloud.io/api/project_badges/measure?project=dperezcabrera_pico-fastapi&metric=alert_status)](https://sonarcloud.io/summary/new_code?id=dperezcabrera_pico-fastapi) [![Duplicated Lines (%)](https://sonarcloud.io/api/project_badges/measure?project=dperezcabrera_pico-fastapi&metric=duplicated_lines_density)](https://sonarcloud.io/summary/new_code?id=dperezcabrera_pico-fastapi) [![Maintainability Rating](https://sonarcloud.io/api/project_badges/measure?project=dperezcabrera_pico-fastapi&metric=sqale_rating)](https://sonarcloud.io/summary/new_code?id=dperezcabrera_pico-fastapi) [![Docs](https://img.shields.io/badge/Docs-pico--fastapi-blue?style=flat&logo=readthedocs&logoColor=white)](https://dperezcabrera.github.io/pico-fastapi/) # Pico-FastAPI **[Pico-FastAPI](https://github.com/dperezcabrera/pico-fastapi)** seamlessly integrates **[Pico-IoC](https://github.com/dperezcabrera/pico-ioc)** with **[FastAPI](https://github.com/fastapi/fastapi)**, bringing true inversion of control and constructor-based dependency injection to one of the fastest and most elegant Python web frameworks. It provides scoped lifecycles, automatic controller registration, and clean architectural boundaries, without global state and without FastAPI’s function-based dependency system. > 🐍 Requires Python 3.11+ > ⚡ Built on FastAPI > ✅ Fully async-compatible > ✅ Real IoC with constructor injection > ✅ Supports singleton, request, session, and websocket scopes With Pico-FastAPI you get the speed, clarity, and async performance of FastAPI, enhanced by a real IoC container for clean, testable, and maintainable applications. --- ## 🎯 Why pico-fastapi FastAPI’s built-in dependency system is function-based, which often ties business logic to the framework. Pico-FastAPI moves dependency resolution into the IoC container, promoting separation of concerns and testability. | Concern | FastAPI Default | pico-fastapi | |----------|-----------------|---------------| | Dependency injection | Function-based | Constructor-based | | Architecture | Framework-driven | Domain-driven | | Testing | Simulate DI calls | Override components in container | | Scopes | Manual or ad-hoc | Automatic (singleton, request, session, websocket) | --- ## 🧱 Core Features - Controller classes with `@controller` - Route decorators: `@get`, `@post`, `@put`, `@delete`, `@patch`, `@websocket` - Constructor injection for controllers and services - Automatic registration into FastAPI - Scoped resolution via middleware for request, session, and websocket - Full Pico-IoC feature set: profiles, overrides, interceptors, cleanup hooks --- ## 📦 Installation ```bash pip install pico-fastapi ``` --- ## 🚀 Quick Example ```python from pico_fastapi import controller, get @controller(prefix="/api") class ApiController: def __init__(self, service: "MyService"): self.service = service @get("/hello") async def hello(self): return {"msg": self.service.greet()} ``` ```python from pico_ioc import component @component class MyService: def greet(self) -> str: return "hello from service" ``` ```python from pico_ioc import init from fastapi import FastAPI container = init( modules=[ "controllers", "services", "pico_fastapi", ] ) app = container.get(FastAPI) ``` --- # 🚀 Quick Example (with pico-boot auto-discovery) ### 1. Controller ```python from pico_fastapi import controller, get @controller(prefix="/api") class ApiController: def __init__(self, service: "MyService"): self.service = service @get("/hello") async def hello(self): return {"msg": self.service.greet()} ``` ### 2. Service ```python from pico_ioc import component @component class MyService: def greet(self) -> str: return "hello from service" ``` ### 3. App Initialization (Using pico-boot) ```python from pico_boot import init from fastapi import FastAPI # No need to declare "pico_fastapi" anymore. # pico-fastapi is auto-discovered via entry points. container = init( modules=[ "controllers", "services", ] ) app = container.get(FastAPI) ``` --- ## 💬 WebSocket Example ```python from pico_fastapi import controller, websocket from fastapi import WebSocket @controller class ChatController: @websocket("/ws") async def chat(self, websocket: WebSocket): await websocket.accept() while True: msg = await websocket.receive_text() await websocket.send_text(f"Echo: {msg}") ``` --- ## 🧪 Testing with Overrides ```python from pico_ioc import init from fastapi import FastAPI from fastapi.testclient import TestClient class FakeService: def greet(self) -> str: return "test" container = init( modules=["controllers", "services", "pico_fastapi"], overrides={"MyService": FakeService()} ) app = container.get(FastAPI) client = TestClient(app) assert client.get("/api/hello").json() == {"msg": "test"} ``` --- ## 📁 Static Files Configuration Example ```python from dataclasses import dataclass from typing import Protocol, runtime_checkable from fastapi import FastAPI from starlette.staticfiles import StaticFiles from pico_ioc import component, configured from pico_fastapi import FastApiConfigurer @configured(target="self", prefix="fastapi", mapping="tree") @dataclass class StaticSettings: static_dir: str = "public" static_url: str = "/static" @component class StaticFilesConfigurer(FastApiConfigurer): priority = -100 def __init__(self, settings: StaticSettings): self.settings = settings def configure_app(self, app: FastAPI) -> None: app.mount(self.settings.static_url, StaticFiles(directory=self.settings.static_dir), name="static") ``` ```python from pico_ioc import init, configuration, YamlTreeSource from fastapi import FastAPI container = init( modules=[ "pico_fastapi", "static_config", ], config=configuration( YamlTreeSource("config.yml") ), ) app = container.get(FastAPI) ``` ```yaml fastapi: title: "My App" version: "1.0.0" debug: true static_dir: "public" static_url: "/assets" ``` --- ## 🔐 JWT Authentication Configuration Example ```python import base64 import json import hmac import hashlib from dataclasses import dataclass from typing import Optional from fastapi import FastAPI, Request from starlette.middleware.base import BaseHTTPMiddleware from pico_ioc import component, configured, PicoContainer from pico_fastapi import FastApiConfigurer def _b64url_decode(data: str) -> bytes: padding = "=" * (-len(data) % 4) return base64.urlsafe_b64decode(data + padding) def _verify_hs256(token: str, secret: str) -> Optional[dict]: parts = token.split(".") if len(parts) != 3: return None header_b64, payload_b64, sig_b64 = parts signing_input = f"{header_b64}.{payload_b64}".encode() expected = hmac.new(secret.encode(), signing_input, hashlib.sha256).digest() try: signature = _b64url_decode(sig_b64) except Exception: return None if not hmac.compare_digest(signature, expected): return None try: payload_json = _b64url_decode(payload_b64) return json.loads(payload_json.decode()) except Exception: return None class JwtMiddleware(BaseHTTPMiddleware): def __init__(self, app, container: PicoContainer, secret: str): super().__init__(app) self.container = container self.secret = secret async def dispatch(self, request: Request, call_next): auth = request.headers.get("Authorization", "") if auth.startswith("Bearer "): token = auth.split(" ", 1)[1] claims = _verify_hs256(token, self.secret) if claims is not None: request.state.jwt_claims = claims response = await call_next(request) return response @dataclass class JwtSettings: secret: str = "changeme" header: str = "Authorization" @component class JwtConfigurer(FastApiConfigurer): priority = 10 def __init__(self, container: PicoContainer, settings: JwtSettings): self.container = container self.settings = settings def configure_app(self, app: FastAPI) -> None: app.add_middleware(JwtMiddleware, container=self.container, secret=self.settings.secret) ``` ```python from pico_ioc import init from fastapi import FastAPI, Request from pico_fastapi import controller, get @controller(prefix="/api") class ProfileController: def __init__(self): pass @get("/me") async def me(self, request: Request): claims = getattr(request.state, "jwt_claims", None) if claims is None: return {"error": "not authenticated"}, 401 return {"sub": claims.get("sub")} container = init( modules=[ "pico_fastapi", "jwt_config", "controllers", ] ) app = container.get(FastAPI) ``` --- ## ⚙️ How It Works * Controller classes are discovered and registered automatically * Each route executes within its own request or websocket scope * All dependencies are resolved via Pico-IoC * Cleanup and teardown occur at FastAPI lifespan No global state and no implicit singletons. --- ## AI Coding Skills Install [Claude Code](https://code.claude.com) or [OpenAI Codex](https://openai.com/index/introducing-codex/) skills for AI-assisted development with pico-fastapi: ```bash curl -sL https://raw.githubusercontent.com/dperezcabrera/pico-skills/main/install.sh | bash -s -- fastapi ``` | Command | Description | |---------|-------------| | `/add-controller` | Add FastAPI controllers with route decorators | | `/add-component` | Add components, factories, interceptors, settings | | `/add-tests` | Generate tests for pico-framework components | All skills: `curl -sL https://raw.githubusercontent.com/dperezcabrera/pico-skills/main/install.sh | bash` See [pico-skills](https://github.com/dperezcabrera/pico-skills) for details. --- ## 📝 License MIT
text/markdown
null
David Perez Cabrera <dperezcabrera@gmail.com>
null
null
MIT License Copyright (c) 2025 David Pérez Cabrera Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
ioc, di, dependency injection, fastapi, inversion of control, spring boot, controller
[ "Development Status :: 4 - Beta", "Framework :: FastAPI", "Topic :: Software Development :: Libraries :: Python Modules", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Typing :: Typed" ]
[]
null
null
>=3.11
[]
[]
[]
[ "pico-ioc>=2.2.0", "fastapi>=0.100", "starlette-session; extra == \"session\"", "uvicorn[standard]; extra == \"run\"" ]
[]
[]
[]
[ "Homepage, https://github.com/dperezcabrera/pico-fastapi", "Repository, https://github.com/dperezcabrera/pico-fastapi", "Issue Tracker, https://github.com/dperezcabrera/pico-fastapi/issues", "Documentation, https://dperezcabrera.github.io/pico-fastapi/", "Changelog, https://github.com/dperezcabrera/pico-fastapi/blob/main/CHANGELOG.md" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:42:20.630904
pico_fastapi-0.3.0.tar.gz
81,068
d3/1e/662d3df4e29f8c75e00e023a3b199f1e96dd6b498b80d82d8a6d89de887e/pico_fastapi-0.3.0.tar.gz
source
sdist
null
false
90dbd775b62657571d3b3cd218f176ff
48fd659c735fc7a6ac88ac1e43ae295880a5109ee6a49ef9f0617db6eb939f78
d31e662d3df4e29f8c75e00e023a3b199f1e96dd6b498b80d82d8a6d89de887e
null
[ "LICENSE" ]
289
2.4
pico-sqlalchemy
0.3.0
Pico-ioc integration for SQLAlchemy. Adds Spring-style transactional support, configuration, and helpers.
# 📦 pico-sqlalchemy [![PyPI](https://img.shields.io/pypi/v/pico-sqlalchemy.svg)](https://pypi.org/project/pico-sqlalchemy/) [![Ask DeepWiki](https://deepwiki.com/badge.svg)](https://deepwiki.com/dperezcabrera/pico-sqlalchemy) [![License: MIT](https://img.shields.io/badge/License-MIT-blue.svg)](https://opensource.org/licenses/MIT) ![CI (tox matrix)](https://github.com/dperezcabrera/pico-sqlalchemy/actions/workflows/ci.yml/badge.svg) [![codecov](https://codecov.io/gh/dperezcabrera/pico-sqlalchemy/branch/main/graph/badge.svg)](https://codecov.io/gh/dperezcabrera/pico-sqlalchemy) [![Quality Gate Status](https://sonarcloud.io/api/project_badges/measure?project=dperezcabrera_pico-sqlalchemy\&metric=alert_status)](https://sonarcloud.io/summary/new_code?id=dperezcabrera_pico-sqlalchemy) [![Duplicated Lines (%)](https://sonarcloud.io/api/project_badges/measure?project=dperezcabrera_pico-sqlalchemy\&metric=duplicated_lines_density)](https://sonarcloud.io/summary/new_code?id=dperezcabrera_pico-sqlalchemy) [![Maintainability Rating](https://sonarcloud.io/api/project_badges/measure?project=dperezcabrera_pico-sqlalchemy\&metric=sqale_rating)](https://sonarcloud.io/summary/new_code?id=dperezcabrera_pico-sqlalchemy) [![Docs](https://img.shields.io/badge/Docs-pico--sqlalchemy-blue?style=flat&logo=readthedocs&logoColor=white)](https://dperezcabrera.github.io/pico-sqlalchemy/) # Pico-SQLAlchemy **Pico-SQLAlchemy** integrates **[Pico-IoC](https://github.com/dperezcabrera/pico-ioc)** with **SQLAlchemy**, providing a true inversion of control persistence layer with **Spring Data-style** declarative features. It brings constructor-based dependency injection, **implicit transaction management**, and powerful **declarative queries** using pure Python and SQLAlchemy’s Async ORM. > 🐍 **Requires Python 3.11+** > 🚀 **Async-Native:** Built entirely on `AsyncSession` and `create_async_engine`. > ✨ **Zero-Boilerplate:** Repositories are transactional by default. > 🔍 **Declarative Queries:** Define SQL or expressions in decorators; the library executes them for you. --- ## 🎯 Why pico-sqlalchemy? Most Python apps suffer from manual session handling (`async with session...`), scattered transaction logic, and verbose repository patterns. **Pico-SQLAlchemy** solves this by offering: | Feature | SQLAlchemy Default | pico-sqlalchemy | | :--- | :--- | :--- | | **Transactions** | Manual `commit()` / `rollback()` | **Implicit** (Auto-managed) | | **Repositories** | DIY Classes | **`@repository`** (Transactional by default) | | **Queries** | Manual implementation | **`@query`** (Declarative execution) | | **Injection** | None / Global variables | **Constructor Injection** (IoC) | | **Pagination** | Manual calculation | **Automatic** (`PageRequest` / `Page`) | --- ## 🧱 Core Features * **Implicit Transactions:** Methods inside `@repository` are automatically **Read-Write** transactional. * **Declarative Queries:** Use `@query` to run SQL or Expressions automatically (defaults to **Read-Only**). * **AOP-Based Propagation:** `REQUIRED`, `REQUIRES_NEW`, `MANDATORY`, `NEVER`, etc. * **Session Lifecycle:** Centralized `SessionManager` handles engine creation and cleanup. * **Pagination:** Built-in support for paged results via `@query(paged=True)`. --- ## 📦 Installation ```bash pip install pico-sqlalchemy ``` You will also need an async database driver: ```bash pip install aiosqlite # for SQLite pip install asyncpg # for PostgreSQL ``` ----- ## 🚀 Quick Example ### 1\. Define Model ```python from sqlalchemy import Integer, String from pico_sqlalchemy import AppBase, Mapped, mapped_column class User(AppBase): __tablename__ = "users" id: Mapped[int] = mapped_column(Integer, primary_key=True) username: Mapped[str] = mapped_column(String(50)) ``` ### 2\. Define Repository (The "Magic" Part) Notice we don't need `@transactional` here. * `save`: Automatically runs in a **Read-Write** transaction. * `find_by_name`: Automatically runs in a **Read-Only** transaction and executes the query logic. <!-- end list --> ```python from pico_sqlalchemy import repository, query, SessionManager, get_session @repository(entity=User) class UserRepository: def __init__(self, manager: SessionManager): self.manager = manager # IMPLICIT: Read-Write Transaction async def save(self, user: User) -> User: session = get_session(self.manager) session.add(user) return user # DECLARATIVE: Read-Only Transaction + Auto-Execution @query(expr="username = :username", unique=True) async def find_by_name(self, username: str) -> User | None: ... # Body is ignored; the library executes the query ``` ### 3\. Define Service Use `@transactional` here to define business logic boundaries. ```python from pico_ioc import component from pico_sqlalchemy import transactional @component class UserService: def __init__(self, repo: UserRepository): self.repo = repo @transactional async def create(self, name: str) -> User: # 1. Check existence (Read-Only tx from repo) existing = await self.repo.find_by_name(name) if existing: raise ValueError("User exists") # 2. Save new user (Joins current transaction) return await self.repo.save(User(username=name)) ``` ### 4\. Run it ```python import asyncio from pico_ioc import init, configuration, DictSource config = configuration(DictSource({ "database": { "url": "sqlite+aiosqlite:///:memory:", "echo": False } })) async def main(): container = init(modules=["pico_sqlalchemy", "__main__"], config=config) service = await container.aget(UserService) user = await service.create("alice") print(f"Created: {user.id}") await container.cleanup_all_async() if __name__ == "__main__": asyncio.run(main()) ``` ----- ## ⚡ Transaction Hierarchy & Rules Pico-SQLAlchemy applies a "Best Effort" strategy to determine transaction configuration. The priority order (highest wins) is: | Priority | Decorator | Default Mode | Use Case | | :--- | :--- | :--- | :--- | | **1 (High)** | **`@transactional(...)`** | Explicit Config | Overriding defaults, Service layer logic. | | **2** | **`@query(...)`** | **Read-Only** | Efficient data fetching. | | **3 (Base)** | **`@repository`** | **Read-Write** | Default for CRUD (saves, updates, deletes). | ### Example Scenarios 1. **Plain Method in Repository:** ```python async def update_user(self): ... ``` 👉 **Result:** Active Read-Write Transaction (Implicit from `@repository`). 2. **Query Method:** ```python @query("SELECT ...") async def get_data(self): ... ``` 👉 **Result:** Active Read-Only Transaction (Implicit from `@query`). 3. **Manual Override:** ```python @transactional(read_only=True) async def complex_report(self): ... ``` 👉 **Result:** Active Read-Only Transaction (Explicit override). ----- ## 🔍 Declarative Queries in Depth The `@query` decorator eliminates boilerplate for common fetches. ### Expression Mode (`expr`) Requires `@repository(entity=Model)`. Injects the expression into a `SELECT * FROM table WHERE ...`. ```python @query(expr="age > :min_age", unique=False) async def find_adults(self, min_age: int) -> list[User]: ... ``` ### SQL Mode (`sql`) Executes raw SQL. Useful for complex joins or specific DTOs. ```python @query(sql="SELECT count(*) as cnt FROM users") async def count_users(self) -> int: ... ``` ### Automatic Pagination Just add `paged=True` and a `page: PageRequest` parameter. ```python from pico_sqlalchemy import Page, PageRequest @query(expr="active = true", paged=True) async def find_active(self, page: PageRequest) -> Page[User]: ... ``` ----- ## 🧪 Testing Testing is simple because you can override the configuration or the components easily using Pico-IoC. ```python @pytest.mark.asyncio async def test_service(): # Setup container with in-memory DB container = ... service = await container.aget(UserService) user = await service.create("test") assert user.id is not None ``` ----- ## 💡 Architecture Overview ``` ┌─────────────────────────────┐ │ Your App │ └──────────────┬──────────────┘ │ Constructor Injection │ ┌──────────────▼───────────────┐ │ Pico-IoC │ └──────────────┬───────────────┘ │ ┌──────────────▼───────────────┐ │ pico-sqlalchemy │ │ 1. Implicit Repo Transactions│ │ 2. Declarative @query │ │ 3. Explicit @transactional │ └──────────────┬───────────────┘ │ SQLAlchemy (Async ORM) ``` ----- ## AI Coding Skills Install [Claude Code](https://code.claude.com) or [OpenAI Codex](https://openai.com/index/introducing-codex/) skills for AI-assisted development with pico-sqlalchemy: ```bash curl -sL https://raw.githubusercontent.com/dperezcabrera/pico-skills/main/install.sh | bash -s -- sqlalchemy ``` | Command | Description | |---------|-------------| | `/add-repository` | Add SQLAlchemy entities and repositories with transactions | | `/add-component` | Add components, factories, interceptors, settings | | `/add-tests` | Generate tests for pico-framework components | All skills: `curl -sL https://raw.githubusercontent.com/dperezcabrera/pico-skills/main/install.sh | bash` See [pico-skills](https://github.com/dperezcabrera/pico-skills) for details. --- ## 📝 License MIT
text/markdown
null
David Perez Cabrera <dperezcabrera@gmail.com>
null
null
MIT License Copyright (c) 2025 David Perez Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
ioc, di, dependency injection, sqlalchemy, transaction, orm, inversion of control, asyncio
[ "Development Status :: 4 - Beta", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: Database", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Typing :: Typed" ]
[]
null
null
>=3.11
[]
[]
[]
[ "pico-ioc>=2.2.0", "sqlalchemy>=2.0", "asyncpg>=0.29.0; extra == \"async\"", "pytest>=8; extra == \"test\"", "pytest-asyncio>=0.23.5; extra == \"test\"", "pytest-cov>=5; extra == \"test\"" ]
[]
[]
[]
[ "Homepage, https://github.com/dperezcabrera/pico-sqlalchemy", "Repository, https://github.com/dperezcabrera/pico-sqlalchemy", "Issue Tracker, https://github.com/dperezcabrera/pico-sqlalchemy/issues", "Documentation, https://dperezcabrera.github.io/pico-sqlalchemy/", "Changelog, https://github.com/dperezcabrera/pico-sqlalchemy/blob/main/CHANGELOG.md" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:42:14.860108
pico_sqlalchemy-0.3.0.tar.gz
73,550
ef/1d/36c19ad583385f97a4514c264bf1a264826bd68149cc5b18f5ab00e91eb3/pico_sqlalchemy-0.3.0.tar.gz
source
sdist
null
false
f942bfe1091b5502ac1c95019c2911f4
ebb9094fb67cbeaa9310c0ef8f80fec983e16485e4cc2dd747d1a41069b89999
ef1d36c19ad583385f97a4514c264bf1a264826bd68149cc5b18f5ab00e91eb3
null
[ "LICENSE" ]
257
2.1
outerbounds
0.12.14
More Data Science, Less Administration
# Outerbounds Main package for the Outerbounds platform.
text/markdown
Outerbounds, Inc.
null
null
null
Proprietary
data science, machine learning, MLOps
[ "Development Status :: 4 - Beta", "License :: Other/Proprietary License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12" ]
[]
null
null
<4.0,>=3.7
[]
[]
[]
[ "azure-identity<2.0.0,>=1.15.0; extra == \"azure\"", "azure-keyvault-secrets<5.0.0,>=4.7.0; extra == \"azure\"", "azure-storage-blob<13.0.0,>=12.9.0; extra == \"azure\"", "boto3", "google-api-core<3.0.0,>=2.16.1; extra == \"gcp\"", "google-auth<3.0.0,>=2.27.0; extra == \"gcp\"", "google-cloud-secret-manager<3.0.0,>=2.20.0; extra == \"gcp\"", "google-cloud-storage<3.0.0,>=2.14.0; extra == \"gcp\"", "metaflow-torchrun>=0.2.1", "metaflow_checkpoint==0.2.10", "ob-metaflow==2.19.19.2", "ob-metaflow-extensions==1.6.11", "ob-metaflow-stubs==6.0.12.14", "ob-project-utils>=0.2.16", "opentelemetry-distro>=0.41b0; extra == \"otel\"", "opentelemetry-exporter-otlp-proto-http>=1.20.0; extra == \"otel\"", "opentelemetry-instrumentation-requests>=0.41b0; extra == \"otel\"", "packaging<25.0,>=24.0; extra == \"gcp\"" ]
[]
[]
[]
[ "Documentation, https://docs.metaflow.org" ]
twine/6.1.0 CPython/3.12.9
2026-02-20T21:42:05.244140
outerbounds-0.12.14-py3-none-any.whl
337,852
6b/14/4e7440e090e7d041ca1d6aefe13814cd343861615b59dd473e3224bd0e3d/outerbounds-0.12.14-py3-none-any.whl
py3
bdist_wheel
null
false
243fe6d40b450e5a70f81af87f2892b6
b7ceaeb0f86da812e6317989e4a3e812242267b813964af63f6a74f479bd0f91
6b144e7440e090e7d041ca1d6aefe13814cd343861615b59dd473e3224bd0e3d
null
[]
3,334
2.4
relace-mcp
0.2.5a2
Unofficial Relace MCP Server - Fast code merging via Relace API
<p align="right"> <strong>English</strong> | <a href="README.zh-CN.md">简体中文</a> </p> # Unofficial Relace MCP Server [![PyPI](https://img.shields.io/pypi/v/relace-mcp.svg?style=flat-square)](https://pypi.org/project/relace-mcp/) [![Python 3.11+](https://img.shields.io/badge/python-3.11+-blue.svg?style=flat-square)](https://www.python.org/downloads/) [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg?style=flat-square)](LICENSE) ![100% AI-Generated](https://img.shields.io/badge/100%25%20AI-Generated-ff69b4.svg?style=flat-square) [![OpenSSF Scorecard](https://img.shields.io/ossf-scorecard/github.com/possible055/relace-mcp?style=flat-square)](https://scorecard.dev/viewer/?uri=github.com/possible055/relace-mcp) > **Unofficial** — Personal project, not affiliated with Relace. > > **Built with AI** — Developed entirely with AI assistance (Antigravity, Codex, Cursor, Github Copilot, Windsurf). MCP server providing AI-powered code editing and intelligent codebase exploration tools. | Without | With `agentic_search` + `fast_apply` | |:--------|:-------------------------------------| | Manual grep, misses related files | Ask naturally, get precise locations | | Edits break imports elsewhere | Traces imports and call chains | | Full rewrites burn tokens | Describe changes, no line numbers | | Line number errors corrupt code | 10,000+ tokens/sec merging | ## Features - **Fast Apply** — Apply code edits at 10,000+ tokens/sec via Relace API - **Agentic Search** — Agentic codebase exploration with natural language queries - **Agentic Retrieval** — Two-stage semantic + agentic code retrieval (requires `MCP_SEARCH_RETRIEVAL=1`) - **Cloud Search** — Semantic code search over cloud-synced repositories ## Quick Start **Prerequisites:** [uv](https://docs.astral.sh/uv/), [git](https://git-scm.com/), [ripgrep](https://github.com/BurntSushi/ripgrep) (recommended) Using Relace (default) or `RELACE_CLOUD_TOOLS=1`: get your API key from [Relace Dashboard](https://app.relace.ai/settings/billing), then add to your MCP client: <details> <summary><strong>Cursor</strong></summary> `~/.cursor/mcp.json` ```json { "mcpServers": { "relace": { "command": "uv", "args": ["tool", "run", "relace-mcp"], "env": { "RELACE_API_KEY": "rlc-your-api-key", "MCP_BASE_DIR": "/absolute/path/to/your/project" } } } } ``` </details> <details> <summary><strong>Claude Code</strong></summary> ```bash claude mcp add relace \ --env RELACE_API_KEY=rlc-your-api-key \ --env MCP_BASE_DIR=/absolute/path/to/your/project \ -- uv tool run relace-mcp ``` </details> <details> <summary><strong>Windsurf</strong></summary> `~/.codeium/windsurf/mcp_config.json` ```json { "mcpServers": { "relace": { "command": "uv", "args": ["tool", "run", "relace-mcp"], "env": { "RELACE_API_KEY": "rlc-your-api-key", "MCP_BASE_DIR": "/absolute/path/to/your/project" } } } } ``` </details> <details> <summary><strong>VS Code</strong></summary> `.vscode/mcp.json` ```json { "mcp": { "servers": { "relace": { "type": "stdio", "command": "uv", "args": ["tool", "run", "relace-mcp"], "env": { "RELACE_API_KEY": "rlc-your-api-key", "MCP_BASE_DIR": "${workspaceFolder}" } } } } } ``` </details> <details> <summary><strong>Codex CLI</strong></summary> `~/.codex/config.toml` ```toml [mcp_servers.relace] command = "uv" args = ["tool", "run", "relace-mcp"] [mcp_servers.relace.env] RELACE_API_KEY = "rlc-your-api-key" MCP_BASE_DIR = "/absolute/path/to/your/project" ``` </details> ## Configuration | Variable | Required | Description | |----------|----------|-------------| | `RELACE_API_KEY` | ✅* | API key from [Relace Dashboard](https://app.relace.ai/settings/billing) | | `RELACE_CLOUD_TOOLS` | ❌ | Set to `1` to enable cloud tools | | `MCP_SEARCH_RETRIEVAL` | ❌ | Set to `1` to enable `agentic_retrieval` tool | | `SEARCH_LSP_TOOLS` | ❌ | LSP tools: `1` (all on), `auto` (detect installed servers), `0` (off, default) | | `MCP_BASE_DIR` | ❌ | Project root (auto-detected via MCP Roots → Git → CWD) | | `MCP_LOGGING` | ❌ | File logging: `off` (default), `safe`, `full` | | `MCP_DOTENV_PATH` | ❌ | Path to `.env` file for centralized config | `*` Optional if **both**: (1) `APPLY_PROVIDER` and `SEARCH_PROVIDER` are non-Relace providers, and (2) `RELACE_CLOUD_TOOLS=false`. For `.env` usage, encoding settings, custom LLM providers, and more, see [docs/advanced.md](docs/advanced.md). ## Tools Core tools (`fast_apply`, `agentic_search`) are always available. Cloud tools require `RELACE_CLOUD_TOOLS=1`. `agentic_retrieval` requires `MCP_SEARCH_RETRIEVAL=1`. For detailed parameters, see [docs/tools.md](docs/tools.md). ## Language Support LSP tools use external language servers installed on your system. | Language | Language Server | Install Command | |----------|-----------------|-----------------| | Python | basedpyright | (bundled) | | TypeScript/JS | typescript-language-server | `npm i -g typescript-language-server typescript` | | Go | gopls | `go install golang.org/x/tools/gopls@latest` | | Rust | rust-analyzer | `rustup component add rust-analyzer` | ## Dashboard Real-time terminal UI for monitoring operations. ```bash pip install relace-mcp[tools] relogs ``` For detailed usage, see [docs/dashboard.md](docs/dashboard.md). ## Benchmark Evaluate `agentic_search` performance using the [Loc-Bench](https://huggingface.co/datasets/IvanaXu/LocAgent) code localization dataset. ```bash # Install benchmark dependencies pip install relace-mcp[benchmark] # Build dataset from Hugging Face uv run python -m benchmark.cli.build_locbench --output artifacts/data/raw/locbench_v1.jsonl # Run evaluation uv run python -m benchmark.cli.run --dataset artifacts/data/raw/locbench_v1.jsonl --limit 20 ``` For grid search, analysis tools, and metrics interpretation, see [docs/benchmark.md](docs/benchmark.md). ## Platform Support | Platform | Status | Notes | |----------|--------|-------| | Linux | ✅ Fully supported | Primary development platform | | macOS | ✅ Fully supported | All features available | | Windows | ⚠️ Partial | `bash` tool unavailable; use WSL for full functionality | ## Troubleshooting | Error | Solution | |-------|----------| | `RELACE_API_KEY is not set` | Set the key in your environment or MCP config | | `NEEDS_MORE_CONTEXT` | Include 1–3 anchor lines before/after target block | | `FILE_TOO_LARGE` | File exceeds 10MB; split file | | `ENCODING_ERROR` | Set `RELACE_DEFAULT_ENCODING` explicitly | | `AUTH_ERROR` | Verify API key is valid and not expired | | `RATE_LIMIT` | Too many requests; wait and retry | | `CONNECTION_TIMEOUT` | Check network connection or increase timeout setting | | `INVALID_PATH` | File path doesn't exist or no permission; verify path and access rights | | `SYNTAX_ERROR` | Invalid edit_snippet format; ensure placeholder syntax is correct | | `NO_MATCH_FOUND` | No search results; try broader query or run `cloud_sync` first | | `CLOUD_NOT_SYNCED` | Repository not synced to Relace Cloud; run `cloud_sync` first | | `CONFLICT_DETECTED` | Edit conflict; file was modified, re-read before editing | ## Development ```bash git clone https://github.com/possible055/relace-mcp.git cd relace-mcp uv sync --extra dev uv run pytest ``` ## License MIT
text/markdown
possible055
null
null
null
MIT
relace, instant-apply, model-context-protocol, mcp-server
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Software Development :: Libraries :: Python Modules", "Framework :: FastAPI", "Environment :: Plugins" ]
[]
null
null
>=3.11
[]
[]
[]
[ "charset-normalizer>=3.0.0", "fastmcp<3,>=2.0.0", "httpx>=0.28.0", "openai>=1.0.0", "pathspec>=0.12.0", "pydantic<3,>=2.0.0", "platformdirs>=4.0.0", "psutil>=5.9.0", "python-dotenv>=1.0.0", "PyYAML>=6.0.0", "tenacity>=8.0.0", "basedpyright>=1.20.0; extra == \"pyright\"", "ty>=0.0.17; extra == \"dev\"", "pytest>=8.4.1; extra == \"dev\"", "pytest-asyncio>=1.3.0; extra == \"dev\"", "pytest-cov>=6.0.0; extra == \"dev\"", "pre-commit>=4.5.1; extra == \"dev\"", "mypy>=1.19.1; extra == \"dev\"", "interrogate>=1.7.0; extra == \"dev\"", "dead>=1.5.2; extra == \"dev\"", "ruff>=0.1.0; extra == \"dev\"", "build>=1.0.0; extra == \"dev\"", "types-PyYAML>=6.0.0; extra == \"dev\"", "bandit[toml]>=1.9.0; extra == \"dev\"", "textual>=0.50.0; extra == \"tools\"", "click>=8.0.0; extra == \"benchmark\"", "rich>=13.0.0; extra == \"benchmark\"", "tree-sitter>=0.24.0; extra == \"benchmark\"", "tree-sitter-python>=0.24.0; extra == \"benchmark\"" ]
[]
[]
[]
[ "Repository, https://github.com/possible055/relace-mcp", "Documentation, https://github.com/possible055/relace-mcp#readme", "Issues, https://github.com/possible055/relace-mcp/issues" ]
twine/6.1.0 CPython/3.12.9
2026-02-20T21:41:50.936143
relace_mcp-0.2.5a2.tar.gz
146,673
06/0e/fd04a909e136f3c8892342a195e37450ba26087f13d6d825716b47080d69/relace_mcp-0.2.5a2.tar.gz
source
sdist
null
false
265be4b4f5b28d3f31e3f42c9a674a93
d1b677c6a575f2132fcc5ecda82ecc9c5112f9e13f661ecff2dfd4628ec56456
060efd04a909e136f3c8892342a195e37450ba26087f13d6d825716b47080d69
null
[ "LICENSE" ]
173
2.4
airbyte-agent-pylon
0.1.2
Airbyte Pylon Connector for AI platforms
# Pylon The Pylon agent connector is a Python package that equips AI agents to interact with Pylon through strongly typed, well-documented tools. It's ready to use directly in your Python app, in an agent framework, or exposed through an MCP. Pylon is a customer support platform that helps B2B companies manage customer interactions across Slack, email, chat widgets, and other channels. This connector provides access to issues, accounts, contacts, teams, tags, users, custom fields, ticket forms, and user roles for customer support analytics and account intelligence insights. ## Example questions The Pylon connector is optimized to handle prompts like these. - List all open issues in Pylon - Show me all accounts in Pylon - List all contacts in Pylon - What teams are configured in my Pylon workspace? - Show me all tags used in Pylon - List all users in my Pylon account - Show me the custom fields configured for issues - List all ticket forms in Pylon - What user roles are available in Pylon? - Show me details for a specific issue - Get details for a specific account - Show me details for a specific contact - What are the most common issue sources this month? - Show me issues assigned to a specific team - Which accounts have the most open issues? - Analyze issue resolution times over the last 30 days - List contacts associated with a specific account ## Unsupported questions The Pylon connector isn't currently able to handle prompts like these. - Delete an issue - Delete an account - Send a message to a customer - Schedule a meeting with a contact ## Installation ```bash uv pip install airbyte-agent-pylon ``` ## Usage Connectors can run in open source or hosted mode. ### Open source In open source mode, you provide API credentials directly to the connector. ```python from airbyte_agent_pylon import PylonConnector from airbyte_agent_pylon.models import PylonAuthConfig connector = PylonConnector( auth_config=PylonAuthConfig( api_token="<Your Pylon API token. Only admin users can create API tokens.>" ) ) @agent.tool_plain # assumes you're using Pydantic AI @PylonConnector.tool_utils async def pylon_execute(entity: str, action: str, params: dict | None = None): return await connector.execute(entity, action, params or {}) ``` ### Hosted In hosted mode, API credentials are stored securely in Airbyte Cloud. You provide your Airbyte credentials instead. If your Airbyte client can access multiple organizations, also set `organization_id`. This example assumes you've already authenticated your connector with Airbyte. See [Authentication](AUTH.md) to learn more about authenticating. If you need a step-by-step guide, see the [hosted execution tutorial](https://docs.airbyte.com/ai-agents/quickstarts/tutorial-hosted). ```python from airbyte_agent_pylon import PylonConnector, AirbyteAuthConfig connector = PylonConnector( auth_config=AirbyteAuthConfig( customer_name="<your_customer_name>", organization_id="<your_organization_id>", # Optional for multi-org clients airbyte_client_id="<your-client-id>", airbyte_client_secret="<your-client-secret>" ) ) @agent.tool_plain # assumes you're using Pydantic AI @PylonConnector.tool_utils async def pylon_execute(entity: str, action: str, params: dict | None = None): return await connector.execute(entity, action, params or {}) ``` ## Full documentation ### Entities and actions This connector supports the following entities and actions. For more details, see this connector's [full reference documentation](REFERENCE.md). | Entity | Actions | |--------|---------| | Issues | [List](./REFERENCE.md#issues-list), [Create](./REFERENCE.md#issues-create), [Get](./REFERENCE.md#issues-get), [Update](./REFERENCE.md#issues-update) | | Messages | [List](./REFERENCE.md#messages-list) | | Issue Notes | [Create](./REFERENCE.md#issue-notes-create) | | Issue Threads | [Create](./REFERENCE.md#issue-threads-create) | | Accounts | [List](./REFERENCE.md#accounts-list), [Create](./REFERENCE.md#accounts-create), [Get](./REFERENCE.md#accounts-get), [Update](./REFERENCE.md#accounts-update) | | Contacts | [List](./REFERENCE.md#contacts-list), [Create](./REFERENCE.md#contacts-create), [Get](./REFERENCE.md#contacts-get), [Update](./REFERENCE.md#contacts-update) | | Teams | [List](./REFERENCE.md#teams-list), [Create](./REFERENCE.md#teams-create), [Get](./REFERENCE.md#teams-get), [Update](./REFERENCE.md#teams-update) | | Tags | [List](./REFERENCE.md#tags-list), [Create](./REFERENCE.md#tags-create), [Get](./REFERENCE.md#tags-get), [Update](./REFERENCE.md#tags-update) | | Users | [List](./REFERENCE.md#users-list), [Get](./REFERENCE.md#users-get) | | Custom Fields | [List](./REFERENCE.md#custom-fields-list), [Get](./REFERENCE.md#custom-fields-get) | | Ticket Forms | [List](./REFERENCE.md#ticket-forms-list) | | User Roles | [List](./REFERENCE.md#user-roles-list) | | Tasks | [Create](./REFERENCE.md#tasks-create), [Update](./REFERENCE.md#tasks-update) | | Projects | [Create](./REFERENCE.md#projects-create), [Update](./REFERENCE.md#projects-update) | | Milestones | [Create](./REFERENCE.md#milestones-create), [Update](./REFERENCE.md#milestones-update) | | Articles | [Create](./REFERENCE.md#articles-create), [Update](./REFERENCE.md#articles-update) | | Collections | [Create](./REFERENCE.md#collections-create) | | Me | [Get](./REFERENCE.md#me-get) | ### Authentication For all authentication options, see the connector's [authentication documentation](AUTH.md). ### Pylon API docs See the official [Pylon API reference](https://docs.usepylon.com/pylon-docs/developer/api/api-reference). ## Version information - **Package version:** 0.1.2 - **Connector version:** 0.1.3 - **Generated with Connector SDK commit SHA:** e9e5b3844f4992b8c672343e5ab34e30da30242c - **Changelog:** [View changelog](https://github.com/airbytehq/airbyte-agent-connectors/blob/main/connectors/pylon/CHANGELOG.md)
text/markdown
null
Airbyte <contact@airbyte.io>
null
null
Elastic-2.0
agent, ai, airbyte, api, connector, data-integration, llm, mcp, pylon
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: Other/Proprietary License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.13", "Topic :: Software Development :: Libraries :: Application Frameworks", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed" ]
[]
null
null
>=3.13
[]
[]
[]
[ "httpx>=0.24.0", "jinja2>=3.0.0", "jsonpath-ng>=1.6.1", "jsonref>=1.1.0", "opentelemetry-api>=1.37.0", "opentelemetry-sdk>=1.37.0", "pydantic>=2.0.0", "python-dotenv>=1.0.0", "pyyaml>=6.0", "segment-analytics-python>=2.2.0" ]
[]
[]
[]
[ "Homepage, https://github.com/airbytehq/airbyte-agent-connectors", "Documentation, https://docs.airbyte.com/ai-agents/", "Repository, https://github.com/airbytehq/airbyte-agent-connectors", "Issues, https://github.com/airbytehq/airbyte-agent-connectors/issues" ]
twine/6.2.0 CPython/3.13.11
2026-02-20T21:41:46.491231
airbyte_agent_pylon-0.1.2.tar.gz
146,264
4b/66/828943a688c5671a5c75217deb80fa323da7c542d0a70efcbdf5dec2f958/airbyte_agent_pylon-0.1.2.tar.gz
source
sdist
null
false
0df7e738ee4444356842284d9c774de5
c2fa517844cfc138f09dce73d3d370c0c8ca4d9123a626569281c80a95bbfecb
4b66828943a688c5671a5c75217deb80fa323da7c542d0a70efcbdf5dec2f958
null
[]
193
2.4
nya-interview
0.1.0
Create cli-based interviews easily
# nya-interview Create cli-based interviews easily See `__main__.py` (run `python -m nya_interview` to see it working) for a source code example.
text/markdown
null
null
null
null
null
null
[]
[]
null
null
>=3.12
[]
[]
[]
[ "nya-scope>=0.2.1", "rich>=14.3.3" ]
[]
[]
[]
[]
uv/0.9.26 {"installer":{"name":"uv","version":"0.9.26","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Arch Linux","version":null,"id":null,"libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}
2026-02-20T21:41:38.562538
nya_interview-0.1.0.tar.gz
24,578
cd/dd/bb09b256c1142afc5a886c5e04cc5d7a6c3f12f219bf14f9c67728fe6880/nya_interview-0.1.0.tar.gz
source
sdist
null
false
910e5347c578f86430a45202bb847ec8
76075a6d6ed5cc0825d150c4be3af0d75e2a7168893c7f8a35ebba0edf2b841b
cdddbb09b256c1142afc5a886c5e04cc5d7a6c3f12f219bf14f9c67728fe6880
null
[]
215
2.4
ob-metaflow-stubs
6.0.12.14
Metaflow Stubs: Stubs for the metaflow package
# Metaflow Stubs This package contains stub files for `metaflow` and thus offers type hints for various editors (such as `VSCode`) and language servers (such as `Pylance`). ## Installation To install Metaflow Stubs in your local environment, you can install from [PyPi](https://pypi.org/project/metaflow-stubs/): ```sh pip install metaflow-stubs ```
text/markdown
Netflix, Outerbounds & the Metaflow Community
help@outerbounds.co
null
null
Apache License 2.0
null
[]
[]
null
null
>=3.7.0
[]
[]
[]
[]
[]
[]
[]
[]
twine/6.1.0 CPython/3.12.9
2026-02-20T21:41:33.064626
ob_metaflow_stubs-6.0.12.14.tar.gz
197,324
34/8e/d1c8580d908caa65785e18b14dc1101b5022e1d393f773cdec12c844630d/ob_metaflow_stubs-6.0.12.14.tar.gz
source
sdist
null
false
e7cdce96facc0c9e58d94bae5ab95746
e547f220ea5eed754b6d772b99208e56e8e7226adfb2d3dbbbb2324c9eb8ac32
348ed1c8580d908caa65785e18b14dc1101b5022e1d393f773cdec12c844630d
null
[]
3,997
2.4
deepagents
0.4.3
General purpose 'deep agent' with sub-agent spawning, todo list capabilities, and mock file system. Built on LangGraph.
# 🧠🤖 Deep Agents [![PyPI - Version](https://img.shields.io/pypi/v/deepagents?label=%20)](https://pypi.org/project/deepagents/#history) [![PyPI - License](https://img.shields.io/pypi/l/deepagents)](https://opensource.org/licenses/MIT) [![PyPI - Downloads](https://img.shields.io/pepy/dt/deepagents)](https://pypistats.org/packages/deepagents) [![Twitter](https://img.shields.io/twitter/url/https/twitter.com/langchain.svg?style=social&label=Follow%20%40LangChain)](https://x.com/langchain) Looking for the JS/TS version? Check out [Deep Agents.js](https://github.com/langchain-ai/deepagentsjs). To help you ship LangChain apps to production faster, check out [LangSmith](https://smith.langchain.com). LangSmith is a unified developer platform for building, testing, and monitoring LLM applications. ## Quick Install ```bash pip install deepagents # or uv add deepagents ``` ## 🤔 What is this? Using an LLM to call tools in a loop is the simplest form of an agent. This architecture, however, can yield agents that are "shallow" and fail to plan and act over longer, more complex tasks. Applications like "Deep Research", "Manus", and "Claude Code" have gotten around this limitation by implementing a combination of four things: a **planning tool**, **sub agents**, access to a **file system**, and a **detailed prompt**. `deepagents` is a Python package that implements these in a general purpose way so that you can easily create a Deep Agent for your application. For a full overview and quickstart of Deep Agents, the best resource is our [docs](https://docs.langchain.com/oss/python/deepagents/overview). **Acknowledgements: This project was primarily inspired by Claude Code, and initially was largely an attempt to see what made Claude Code general purpose, and make it even more so.** ## 📖 Resources - **[Documentation](https://docs.langchain.com/oss/python/deepagents)** — Full documentation - **[API Reference](https://reference.langchain.com/python/deepagents/)** — Full SDK reference documentation - **[Chat LangChain](https://chat.langchain.com)** - Chat interactively with the docs ## 📕 Releases & Versioning See our [Releases](https://docs.langchain.com/oss/python/release-policy) and [Versioning](https://docs.langchain.com/oss/python/versioning) policies. ## 💁 Contributing As an open-source project in a rapidly developing field, we are extremely open to contributions, whether it be in the form of a new feature, improved infrastructure, or better documentation. For detailed information on how to contribute, see the [Contributing Guide](https://docs.langchain.com/oss/python/contributing/overview).
text/markdown
null
null
null
null
MIT
agents, ai, llm, langgraph, langchain, deep-agent, sub-agents, agentic
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Libraries :: Python Modules" ]
[]
null
null
<4.0,>=3.11
[]
[]
[]
[ "langchain-core<2.0.0,>=1.2.10", "langchain<2.0.0,>=1.2.10", "langchain-anthropic<2.0.0,>=1.3.3", "langchain-google-genai<5.0.0,>=4.2.0", "wcmatch" ]
[]
[]
[]
[ "Homepage, https://docs.langchain.com/oss/python/deepagents/overview", "Documentation, https://reference.langchain.com/python/deepagents/", "Repository, https://github.com/langchain-ai/deepagents", "Issues, https://github.com/langchain-ai/deepagents/issues", "Twitter, https://x.com/LangChain", "Slack, https://www.langchain.com/join-community", "Reddit, https://www.reddit.com/r/LangChain/" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:40:45.467976
deepagents-0.4.3.tar.gz
83,210
b4/30/5bba09d1c196a9e6e2e3a3406cd131bdf01e84ec67c4b6233f68a903978f/deepagents-0.4.3.tar.gz
source
sdist
null
false
11457e3c5a50c1f840cf1a098df826f9
88033c616c5ea481f2620dbb2d05533bc8fdcd48f376d713f9dba49a8157b6f8
b4305bba09d1c196a9e6e2e3a3406cd131bdf01e84ec67c4b6233f68a903978f
null
[]
4,739
2.1
ob-metaflow-extensions
1.6.11
Outerbounds Platform Extensions for Metaflow
# Outerbounds platform package This package installs client side packages for Outerbounds platform. See Outerbounds platform documentation and [Metaflow documentation](https://metaflow.org/) for more info on how to use it.
text/markdown
Outerbounds, Inc.
null
null
null
Commercial
null
[]
[]
null
null
null
[]
[]
[]
[]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:39:32.329903
ob_metaflow_extensions-1.6.11.tar.gz
216,054
de/5d/4bcf806f5b2ab89bb906343e4cd8180873f743934f556d8c91c47f4e91a5/ob_metaflow_extensions-1.6.11.tar.gz
source
sdist
null
false
e2de0aeb6f0da5f3ddfc534178cd3ad4
ab709460ebcbc95276c50cc50c17b6a0f059efa0c678af33f72f3559e844013e
de5d4bcf806f5b2ab89bb906343e4cd8180873f743934f556d8c91c47f4e91a5
null
[]
3,514
2.4
earthdata-varinfo
4.1.0
A package for parsing Earth Observation science granule structure and extracting relations between science variables and their associated metadata, such as coordinates.
# earthdata-varinfo A Python package developed as part of the NASA Earth Observing System Data and Information System (EOSDIS) for parsing Earth Observation science granule structure and extracting relations between science variables and their associated metadata, such as coordinates. This package also includes the capability to generate variable (UMM-Var) metadata records that are compatible with the NASA EOSDIS Common Metadata Repository ([CMR](https://www.earthdata.nasa.gov/eosdis/science-system-description/eosdis-components/cmr)). For general usage of classes and functions in `earthdata-varinfo`, see: <https://github.com/nasa/earthdata-varinfo/blob/main/docs/earthdata-varinfo.ipynb>. ## Features: ### CFConfig A class that takes a JSON file and retrieves all related configuration based on the supplied mission name and collection shortname. The JSON file is optional, and if not supplied, a `CFConfig` class will be constructed with largely empty attributes. ``` python from varinfo import CFConfig cf_config = CFConfig('ICESat2', 'ATL03', config_file='config/0.0.1/sample_config_0.0.1.json') metadata_attributes = cf_config.get_metadata_attributes('/full/variable/path') ``` ### VarInfo A group of classes that contain metadata attributes for all groups and variables in a single granule, and the relations between all variables within that granule. Current classes include: * VarInfoBase: An abstract base class that contains core logic and methods used by the child classes that parse different sources of granule information. * VarInfoFromDmr: Child class that maps input from a `.dmr` file downloaded from Hyrax in the cloud. This inherits all the methods and logic of VarInfoBase. * VarInfoFromNetCDF4: Child class that maps input directly from a NetCDF-4 file. Thus inherits all the methods and logic of VarInfoBase. ``` python from varinfo import VarInfoFromDmr var_info = VarInfoFromDmr('/path/to/local/file.dmr', config_file='config/0.0.1/sample_config_0.0.1.json') # Retrieve a set of variables with coordinate metadata: var_info.get_science_variables() # Retrieve a set of variables without coordinate metadata: var_info.get_metadata_variables() # Augment a set of desired variables with all variables required to support # the requested set. For example coordinate variables. var_info.get_required_variables({'/path/to/science/variable'}) # Retrieve an ordered list of dimensions associated with all specified variables. var_info.get_required_dimensions({'/path/to/science/variable'}) # Retrieve all spatial dimensions associated with the specified set of science # variables. var_info.get_spatial_dimensions({'/path/to/science/variable'}) ``` The `VarInfoFromDmr` and `VarInfoFromNetCDF4` classes also have an optional argument `short_name`, which can be used upon instantiation to specify the short name of the collection to which the granule belongs. This option is the preferred way to specify a collection short name, and particularly encouraged for use when a granule does not contain the collection short name within its metadata attributes (e.g., ABoVE collections from ORNL). ``` python var_info = VarInfoFromDmr('/path/to/local/file.dmr', short_name='ATL03') ``` Note: as there are now two optional parameters, `short_name` and `config_file`, it is best to ensure that both are specified as named arguments upon instantiation. ### UMM-Var generation `earthdata-varinfo` can generate variable metadata records compatible with the CMR UMM-Var schema: ``` python from varinfo import VarInfoFromNetCDF4 from varinfo.umm_var import export_all_umm_var_to_json, get_all_umm_var # Instantiate a VarInfoFromNetCDF4 object for a local NetCDF-4 file. var_info = VarInfoFromNetCDF4('/path/to/local/file.nc4', short_name='ATL03') # Retrieve a dictionary of UMM-Var JSON records. Keys are the full variable # paths, values are UMM-Var schema-compatible, JSON-serialisable dictionaries. umm_var = get_all_umm_var(var_info) # Write each UMM-Var dictionary to its own JSON file: export_all_umm_var_to_json(list(umm_var.values()), output_dir='local_dir') ``` ### End-to-end UMM-Var generation and publication: ``` python from cmr import CMR_OPS from varinfo.generate_umm_var import generate_collection_umm_var # Defaults to UAT, and not to publish: umm_var_json = generate_collection_umm_var(<UAT collection concept ID>, <authorization header>) # To use a production collection: umm_var_json = generate_collection_umm_var(<Production collection concept ID>, <authorization header>, cmr_env=CMR_OPS) # To generate and publish records for a UAT collection (note the authorization # header must contain a LaunchPad token): umm_var_json = generate_collection_umm_var(<UAT collection concept ID>, <authorization header>, publish=True) # Use a DMR file to generate UMM-Var, defaults to UAT, and not to publish: umm_var_json = generate_collection_umm_var(<UAT collection concept ID>, <authorization header>) # To generate and publish records from a DMR file for a UAT collection # (note the authorization header must contain a LaunchPad token): umm_var_json = generate_collection_umm_var(<UAT collection concept ID>, <authorization header>, publish=True, use_dmr=True) ``` Expected outputs: * `publish=False`, or not specifying a value will result in JSON output containing the UMM-Var JSON for each identified variable. * `publish=True` will return a list of strings. Each string is either the concept ID of a new UMM-Var record, or a string including the full path of a variable that failed to publish and the error messages returned from CMR. Native IDs for generated UMM-Var records will be of format: ``` <collection concept ID>-<variable Name> ``` For variables that are hierarchical, slashes will be converted to underscores, to ensure the native ID is compatible with the CMR API. ## Configuration file schema: The configuration file schema is defined as a JSON schema file in the `config` directory. Each new iteration to the schema should be placed in its own semantically versioned subdirectory, and a sample configuration file should be provided. Additionally, notes on the schema changes should be provided in `config/CHANGELOG.md`. ## Installing ### Using pip Install the latest version of the package from PyPI using pip: ```bash $ pip install earthdata-varinfo ``` ### Other methods: For local development, it is possible to clone the repository and then install the version being developed in editable mode: ```bash $ git clone https://github.com/nasa/earthdata-varinfo $ cd earthdata-varinfo $ pip install -e . ``` ## Contributing Contributions are welcome! For more information see `CONTRIBUTING.md`. ## Developing Development within this repository should occur on a feature branch. Pull Requests (PRs) are created with a target of the `main` branch before being reviewed and merged. Releases are created when a feature branch is merged to `main` and that branch also contains an update to the `VERSION` file. ### Development Setup: Prerequisites: - Python 3.9+, ideally installed in a virtual environment, such as `pyenv` or `conda`. - A local copy of this repository. Set up conda virtual environment: ```bash conda create --name earthdata-varinfo python=3.12 --channel conda-forge \ --override-channels -y conda activate earthdata-varinfo ``` Install dependencies: ```bash $ make develop ``` or ```bash pip install -r requirements.txt -r dev-requirements.txt ``` Run a linter against package code (preferably do this prior to submitting code for a PR review): ```bash $ make lint ``` Run `unittest` suite (run via `pytest`, but written using `unittest` classes): ```bash $ make test ``` Note, the test execution will fail if code coverage of unit tests falls below 95%. This threshold is also used during the GitHub workflow CI/CD. ### pre-commit hooks: This repository uses [pre-commit](https://pre-commit.com/) to enable pre-commit checking the repository for some coding standard best practices. These include: * Removing trailing whitespaces. * Removing blank lines at the end of a file. * JSON files have valid formats. * [ruff](https://github.com/astral-sh/ruff) Python linting checks. * [black](https://black.readthedocs.io/en/stable/index.html) Python code formatting checks. To enable these checks: ```bash # Install pre-commit Python package as part of test requirements: pip install -r dev-requirements.txt # Install the git hook scripts: pre-commit install # (Optional) Run against all files: pre-commit run --all-files ``` When you try to make a new commit locally, `pre-commit` will automatically run. If any of the hooks detect non-compliance (e.g., trailing whitespace), that hook will state it failed, and also try to fix the issue. You will need to review and `git add` the changes before you can make a commit. It is planned to implement additional hooks, possibly including tools such as `mypy`. [pre-commit.ci](pre-commit.ci) is configured such that these same hooks will be automatically run for every pull request. ## Releasing: All CI/CD for this repository is defined in the `.github/workflows` directory: * run_tests.yml - A reusable workflow that runs the unit test suite under a matrix of Python versions. * run_tests_on_pull_requests.yml - Triggered for all PRs against main. It runs the workflow in run_test.yml to ensure all tests pass on the new code. * publish_to_pypi.yml - Triggered either manually or for commits to the main branch that contain changes to the `VERSION` file. The `publish_to_pypi.yml` workflow will: * Run the full unit test suite, to prevent publication of broken code. * Extract the semantic version number from `VERSION`. * Extract the release notes for the most recent version from `CHANGELOG.md`. * Build the package to be published to PyPI. * Publish the package to PyPI. * Publish a GitHub release under the semantic version number, with associated git tag. Before triggering a release, ensure the `VERSION` and `CHANGELOG.md` files are updated accordingly. ## Get in touch: You can reach out to the maintainers of this repository via email: * david.p.auty@nasa.gov * owen.m.littlejohns@nasa.gov
text/markdown
NASA EOSDIS SDPS Data Services Team
owen.m.littlejohns@nasa.gov
null
null
License :: OSI Approved :: Apache Software License
null
[ "Programming Language :: Python", "Programming Language :: Python :: 3", "Operating System :: OS Independent" ]
[]
https://github.com/nasa/earthdata-varinfo
null
>=3.9
[]
[]
[]
[ "netCDF4>=1.7.2", "numpy<2.3,>=1.24.2", "python-cmr~=0.12.0", "requests~=2.31.0", "urllib3~=2.6.1", "ipython~=8.18.1; extra == \"dev\"", "jsonschema~=4.23.0; extra == \"dev\"", "pre-commit~=4.2.0; extra == \"dev\"", "pycodestyle~=2.12.1; extra == \"dev\"", "pylint~=3.3.6; extra == \"dev\"", "pytest~=8.3.5; extra == \"dev\"", "pytest-cov~=6.0.0; extra == \"dev\"" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:38:52.336854
earthdata_varinfo-4.1.0.tar.gz
38,460
89/45/0a26eb30aff40e5930a621170c5c50033ce37e8c73da602582647f828594/earthdata_varinfo-4.1.0.tar.gz
source
sdist
null
false
31997a67e1a051a1e643c679fe5cc227
a34bb8f657b44cc1c335f9da24158e267009be2a3555f1f440b7d6fb80e5c995
89450a26eb30aff40e5930a621170c5c50033ce37e8c73da602582647f828594
null
[ "LICENSE" ]
209
2.4
apache-airflow-ctl
0.1.2rc1
Apache Airflow command line tool for communicating with an Apache Airflow, using the API.
<!-- Licensed to the Apache Software Foundation (ASF) under one or more contributor license agreements. See the NOTICE file distributed with this work for additional information regarding copyright ownership. The ASF licenses this file to you under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. --> # airflowctl A command-line tool for interacting with Apache Airflow instances through the Airflow REST API. It offers a convenient interface for performing common operations remotely without direct access to the Airflow scheduler or webserver. ## Features - Communicates with Airflow instances through the REST API - Supports authentication using Airflow API tokens - Executes commands against remote Airflow deployments - Provides intuitive command organization with group-based structure - Includes detailed help documentation for all commands ## Requirements - Python 3.10 or later (compatible with Python >= 3.10 and < 3.13) - Network access to an Apache Airflow instance with REST API enabled - Keyring backend installed in operating system for secure token storage ## Usage Access the tool from your terminal: ### Command Line ```bash airflowctl --help ``` ## Contributing Want to help improve Apache Airflow? Check out our [contributing documentation](https://github.com/apache/airflow/blob/main/contributing-docs/README.rst). ### Additional Contribution Guidelines - Please ensure API is running while doing development testing. - There are two ways to have a CLI command, - Auto Generated Commands - Implemented Commands #### Auto Generated Commands Auto generation of commands directly from operations methods under `airflow-ctl/src/airflowctl/api/operations.py`. Whenever operation is mapped with proper datamodel and response model, it will be automatically added to the command. You can check each command with `airflowctl <command> --help` to see the available options. #### Implemented Commands Implemented commands are the ones which are not auto generated and need to be implemented manually. You can check the implemented commands under `airflow-ctl/src/airflowctl/clt/commands/`.
text/markdown
null
null
null
null
null
null
[ "Framework :: Apache Airflow" ]
[]
null
null
!=3.14,>=3.10
[]
[]
[]
[ "argcomplete>=1.10", "httpx>=0.27.0", "keyring>=25.7.0", "lazy-object-proxy>=1.2.0", "methodtools>=0.4.7", "platformdirs>=4.3.6", "pydantic>=2.11.0", "rich-argparse>=1.0.0", "structlog>=25.4.0", "tabulate>=0.9.0", "uuid6>=2024.7.10", "keyrings-alt>=5.0.2; extra == \"dev\"" ]
[]
[]
[]
[ "Bug Tracker, https://github.com/apache/airflow/issues", "Documentation, https://airflow.staged.apache.org/docs/apache-airflow-ctl/stable/index.html", "Downloads, https://archive.apache.org/dist/airflow/airflow-ctl/", "Homepage, https://airflow.staged.apache.org/", "Release Notes, https://airflow.staged.apache.org/docs/apache-airflow-ctl/stable/changelog.html", "Slack Chat, https://s.apache.org/airflow-slack", "Source Code, https://github.com/apache/airflow", "LinkedIn, https://www.linkedin.com/company/apache-airflow/", "Mastodon, https://fosstodon.org/@airflow", "Bluesky, https://bsky.app/profile/apache-airflow.bsky.social", "YouTube, https://www.youtube.com/channel/UCSXwxpWZQ7XZ1WL3wqevChA/" ]
twine/6.2.0 CPython/3.11.9
2026-02-20T21:38:27.823848
apache_airflow_ctl-0.1.2rc1.tar.gz
158,544
57/9f/0a188908ca17a5a98c322b5b06f9b1378f808607ed5eb1470bf6b3c85f6e/apache_airflow_ctl-0.1.2rc1.tar.gz
source
sdist
null
false
2c8be1e839b048f5056b59aa63f34ec4
33c6fafad5e93b167c1c9b34038fb0042aba8eeb64ec1bbd3169df25f3a2635f
579f0a188908ca17a5a98c322b5b06f9b1378f808607ed5eb1470bf6b3c85f6e
Apache-2.0
[ "LICENSE", "NOTICE" ]
184
2.4
python-ember-mug
1.3.1
Python Library for Ember Mugs.
# Python Ember Mug [![pypi](https://img.shields.io/pypi/v/python-ember-mug.svg)](https://pypi.org/project/python-ember-mug/) [![python](https://img.shields.io/pypi/pyversions/python-ember-mug.svg)](https://pypi.org/project/python-ember-mug/) [![Build Status](https://github.com/sopelj/python-ember-mug/actions/workflows/tests.yml/badge.svg)](https://github.com/sopelj/python-ember-mug/actions/workflows/tests.yml) [![codecov](https://codecov.io/gh/sopelj/python-ember-mug/graph/badge.svg?token=2Lw2iVjKsG)](https://codecov.io/gh/sopelj/python-ember-mug) ![Project Maintenance](https://img.shields.io/maintenance/yes/2026.svg) [![Maintainer](https://img.shields.io/badge/maintainer-%40sopelj-blue.svg)](https://github.com/sopelj) [![License](https://img.shields.io/github/license/sopelj/python-ember-mug.svg)](https://github.com/sopelj/python-ember-mug/blob/main/LICENSE) [![pre-commit](https://img.shields.io/badge/pre--commit-enabled-brightgreen)](https://github.com/pre-commit/pre-commit) Python Library for interacting with Ember Mugs, Cups, and Travel Mugs via Bluetooth * [📘 Documentation](https://sopelj.github.io/python-ember-mug) * [💻 GitHub](https://github.com/sopelj/python-ember-mug) * [🐍 PyPI](https://pypi.org/project/python-ember-mug/) ## Summary This is an *unofficial* library to attempt to interact with Ember Mugs via Bluetooth. This was created for use with my [Home Assistant integration](https://github.com/sopelj/hass-ember-mug-component), but could be useful separately and has a simple CLI interface too. All known Ember Mugs, Cups, Tumblers and Travel Mugs have been tested and seem to work well. If I missed one, or you have new feature ideas or issues, please [create an issue](https://github.com/sopelj/python-ember-mug/issues), if it isn't already there, and we'll figure it out. | Device | Tested | |--------------|--------| | Mug | ✓ | | Mug 2 | ✓ | | Cup | ✓ | | Tumbler | ✓ | | Travel Mug | ✓ | | Travel Mug 2 | ✓ | ## Features * Finding devices * Connecting to devices * Reading/Writing most values * Poll for changes Attributes by device: | Attribute | Mug | Cup | Tumbler | Travel Mug | Description | |---------------------|-----|-----|---------|------------|-----------------------------------------------| | Name | R/W | N/A | N/A | R | Name to give device | | LED Colour | R/W | R/W | R/W | N/A | Colour of front LED | | Current Temperature | R | R | R | R | Current temperature of the liquid in the mug | | Target Temperature | R/W | R/W | R/W | R/W | Desired temperature for the liquid | | Temperature Unit | R/W | R/W | R/W | R/W | Internal temperature unit for the app (C/F) | | Liquid Level | R | R | R | R | Approximate level of the liquid in the device | | Volume level | N/A | N/A | N/A | R/W | Volume of the button press beep | | Battery Percent | R | R | R | R | Current battery level | | On Charger | R | R | R | R | Device is on it's charger | > **Note** > Writing may only work if the devices has been set up in the app previously ## Usage ### Python ```python from ember_mug.scanner import find_device, discover_devices from ember_mug.utils import get_model_info_from_advertiser_data from ember_mug.mug import EmberMug # if first time with mug in pairing devices = await discover_devices() # after paired you can simply use device, advertisement = await find_device() model_info = get_model_info_from_advertiser_data(advertisement) mug = EmberMug(device, model_info) await mug.update_all() print(mug.data.formatted) await mug.disconnect() # You can also use connection as a context manager # if you want to ensure connection before starting and cleanup on exit async with mug.connection(): print('Connected.\nFetching Info') await mug.update_all() print(mug.data.formatted) ``` ### CLI Mode It can also be run via command line either directly with `ember-mug --help` or as a module with `python -m ember_mug --help` There are four options with different subsections. You can see them by specifying them before help. eg `ember-mug poll --help` ```bash ember-mug discover # Finds the mug in pairing mode for the first time ember-mug poll # fetches info and keeps listening for notifications ember-mug get name target-temp # Prints name and target temp of mug ember-mug set --name "My mug" --target-temp 56.8 # Sets the name and target temp to specified values ``` Basic options: | Command | Use | |-------------|-----------------------------------------------------------------------------------| | `discover` | Find/List all detected unpaired devices in pairing mode | | `find` | Find *one* already paired devices | | `info` | Connect to *one* device and print its current state | | `poll` | Connect to *one* device and print its current state and keep watching for changes | | `get` | Get the value(s) of one or more attribute(s) by name | | `set` | Set one or more values on the device | Example: <!-- termynal --> ``` $ ember-mug poll Found device: C9:0F:59:D6:33:F9: Ember Ceramic Mug Connecting... Connected. Fetching Info Device Data +--------------+----------------------+ | Device Name | Jesse's Mug | +--------------+----------------------+ | Meta | None | +--------------+----------------------+ | Battery | 64.0% | | | not on charging base | +--------------+----------------------+ | Firmware | None | +--------------+----------------------+ | LED Colour | #ff0fbb | +--------------+----------------------+ | Liquid State | Empty | +--------------+----------------------+ | Liquid Level | 0.00% | +--------------+----------------------+ | Current Temp | 24.50°C | +--------------+----------------------+ | Target Temp | 55.00°C | +--------------+----------------------+ Watching for changes Current Temp changed from "24.50°C" to "25.50°" Battery changed from "64.0%, on charging base" to "65.5%, on charging base" ``` ## Caveats * Since this api is not public, a lot of guesswork and reverse engineering is involved, so it's not perfect. * If the device has not been set up in the app since it was reset, writing is not allowed. I don't know what they set in the app, but it changes something, and it doesn't work without it. * Once that device has been set up in the app, you should ideally forget the device or at least turn off bluetooth whilst using it here, or you will probably get disconnected often * I haven't figured out some attributes like udsk, dsk, location and timezone, but they are not very useful anyway. ## Troubleshooting ##### Systematic timeouts or `le-connection-abort-by-local` If your mug gets stuck in a state where it refuses to connect, you get constant reconnects, timeouts, and/or `le-connection-abort-by-local` messages in the debug logs, you may need to remove your mug via `bluetoothctl remove my-mac-address` and factory reset your device. It should reconnect correctly afterward. You may also need to re-add it to the app in order to make it writable again as well. ### 'Operation failed with ATT error: 0x0e' or another connection error This seems to be caused by the bluetooth adaptor being in some sort of passive mode. I have not yet figured out how to wake it programmatically so sadly, you need to manually open `bluetoothctl` to do so. Please ensure the device is in pairing mode (ie the light is flashing blue or says "PAIR") and run the `bluetoothctl` command. You don't need to type anything. run it and wait until the mug connects. ### Model incorrect or not found I don't have a lot of these devices, so if this library does not correctly identify your device, please open and issue with the advertisement data of your device, so I can update the library to correctly identify it. Thanks! ## Development Install: - [hatch](https://hatch.pypa.io/latest/install/) - [pre-commit](https://pre-commit.com/) ```bash pip install hatch # Use CLI interface hatch run ember-mug --help # Run Tests hatch run test:cov # View docs hatch docs:serve # Lint code pre-commit run --all-files ``` ## Credits This package was created with [Cookiecutter](https://github.com/audreyr/cookiecutter) and the [waynerv/cookiecutter-pypackage](https://github.com/waynerv/cookiecutter-pypackage) project template. ## Notice of Non-Affiliation and Disclaimer This project is not affiliated, associated, authorized, endorsed by, or in any way officially connected with Ember. The name Ember as well as related names, marks, emblems and images are registered trademarks of their respective owners.
text/markdown
null
Jesse Sopel <jesse@sopelj.ca>
null
null
null
null
[ "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14" ]
[]
null
null
>=3.11
[]
[]
[]
[ "bleak-retry-connector>=4.0.2", "bleak<2.2.0,>=1.0.1", "ipython; extra == \"dev\"", "black; extra == \"docs\"", "mkdocs-autorefs; extra == \"docs\"", "mkdocs-gen-files; extra == \"docs\"", "mkdocs-include-markdown-plugin<8.0.0,>=7.0.0; extra == \"docs\"", "mkdocs-literate-nav; extra == \"docs\"", "mkdocs-material-extensions; extra == \"docs\"", "mkdocs-material<10.0.0,>=9.5.44; extra == \"docs\"", "mkdocs>=1.6.1; extra == \"docs\"", "mkdocstrings-python>=1.12.0; extra == \"docs\"", "termynal; extra == \"docs\"", "pytest-asyncio; extra == \"test\"", "pytest-cov; extra == \"test\"", "pytest>=7.2.1; extra == \"test\"" ]
[]
[]
[]
[ "Changelog, https://sopelj.github.io/python-ember-mug/changelog/", "Documentation, https://sopelj.github.io/python-ember-mug/", "Source code, https://github.com/sopelj/python-ember-mug/", "Bug Tracker, https://github.com/sopelj/python-ember-mug/issues" ]
Hatch/1.16.3 cpython/3.12.12 HTTPX/0.28.1
2026-02-20T21:37:52.489872
python_ember_mug-1.3.1-py3-none-any.whl
27,470
de/7a/9d4d7ede39afc3d5d127e3d2c5adca567a3792946420aa34914806d5421b/python_ember_mug-1.3.1-py3-none-any.whl
py3
bdist_wheel
null
false
5f148d08047a13026075953a76fce2d6
67677a0073b88dc068e44b9f96cec97adc989ec740f3626d2513ca5b50742743
de7a9d4d7ede39afc3d5d127e3d2c5adca567a3792946420aa34914806d5421b
MIT
[ "LICENSE" ]
223
2.4
phaseshift
1.0.0
Decompositions and approximations of linear optical unitaries.
[![Paper](https://img.shields.io/badge/Paper-Read%20Article-blue)](https://doi.org/10.1364/JOSAB.577579) # PhaseShift Decomposition and approximation tools for linear optical unitaries. ## Table of Contents - [About this project](#about-this-project) - [Package contents](#package-contents) - [Installation](#installation) - [Usage](#usage) - [Documentation](#documentation) - [Citing this work](#citing-this-work) - [License](#license) - [References](#references) ## About This Project `PhaseShift` is a **Python** package for performing various **decompositions** and **approximations** of unitary matrices into planar arrangements of simple optical components. These tools can be used to design and program **universal multiport interferometers** (UMIs), devices capable of implementing arbitrary linear transformations on multiple optical modes. Such devices have numerous applications in communication, imaging, and information processing. The algorithms implemented in this package cover two main classes of planar UMI architectures: - [Networks of two-mode components](#two-mode-component-networks) - [Sequences of multichannel components](#multichannel-component-sequences) ### Two-mode Component Networks This class of decompositions seeks to express an $N \times N$ unitary matrix as a planar mesh of configurable two-mode unit cells, typically realized using Mach–Zehnder interferometers (MZIs). The first design of this kind was proposed by [Reck *et al.*, 1994](https://doi.org/10.1103/PhysRevLett.73.58), who used a triangular mesh of asymmetric Mach–Zehnder interferometers to implement arbitrary unitary transformations. This design was later improved by [Clements *et al.*, 2016](https://doi.org/10.1364/OPTICA.3.001460), who introduced a more compact rectangular mesh using the same unit cells. [Bell *et al.*, 2021](https://doi.org/10.1063/5.0053421) further compactified the Clements *et al.* design by using a symmetric Mach–Zehnder interferometer as unit cell, which helped reduce the optical depth of the interferometer. <div align="center"> Rectangular mesh of Mach–Zehnder interferometers based on the Clements architecture for 6 modes </div> ![Clements design](https://github.com/polyquantique/phaseshift/raw/main/figures/Clements.svg) ### Multichannel Component Sequences This second class of decompositions aims to express an $N \times N$ unitary matrix as a sequence of configurable phase masks interleaved with a multichannel mixing layer, such as the discrete Fourier transform (DFT). Numerical evidence suggests that using $N+1$ layers of phase masks with any dense mixing layer is enough to result in a universal design [(Saygin *et al.*, 2020)](https://doi.org/10.1103/PhysRevLett.124.010501) [(Zelaya *et al.*, 2024)](https://doi.org/10.1038/s41598-024-60700-8). The first constructive design based on this approach was proposed by [López Pastor *et al.*, 2021](https://doi.org/10.1364/OE.432787) and generates a sequence of $6N + 1$ phase masks to implement a $N \times N$ unitary. We improved this design to reach $4N+1$ and $2N+5$ phase masks [(Girouard *et al.*, 2026)](https://doi.org/10.1364/JOSAB.577579). <div align="center"> Sequence of phase masks interleaved with the discrete Fourier transform mixing layer for 6 modes </div> ![Fourier design](https://github.com/polyquantique/phaseshift/raw/main/figures/Fourier.svg) ## Package Contents `PhaseShift` provides tools to perform **exact decompositions** and **numerical approximations** of unitary matrices. ### Exact Decompositions `PhaseShift` includes four main modules to perform exact decompositions of unitary matrices: - [`clements_interferometer`](src/phaseshift/clements_interferometer.py): Implementation of the algorithm by [Clements *et al.*, 2016](https://doi.org/10.1364/OPTICA.3.001460) to decompose $N \times N$ unitary matrices into a rectangular mesh of $N(N-1)/2$ **asymmetric** Mach–Zehnder interferometers. - [`bell_interferometer`](src/phaseshift/bell_interferometer.py): Implementation of the algorithm by [Bell *et al.*, 2021](https://doi.org/10.1063/5.0053421) to decompose $N \times N$ unitary matrices into a rectangular mesh of $N(N-1)/2$ **symmetric** Mach–Zehnder interferometers. - [`lplm_interferometer`](src/phaseshift/lplm_interferometer.py): Implementation of the algorithm by [López Pastor *et al.*, 2021](https://doi.org/10.1364/OE.432787) to decompose $N \times N$ unitary matrices into a sequence of $6N+1$ phase masks interleaved with the **DFT** matrix. - [`fourier_interferometer`](src/phaseshift/fourier_interferometer.py): Implementation of the **Fourier decomposition** and the **compact Fourier decomposition** [(Girouard *et al.*, 2026)](https://doi.org/10.1364/JOSAB.577579) to decompose $N \times N$ unitary matrices into sequences of $4N+1$ and $2N+5$ phase masks interleaved with **DFT** respectively. ### Optimization Tools In addition to exact decompositions, `PhaseShift` also has an `optimization` subpackage, which contains tools to approximate unitary matrices into a sequence of phase masks interleaved with a chosen mixing layer. The `optimization` subpackage has two modules: - [`fourier_optimizer`](src/phaseshift/optimization/fourier_optimizer.py): Uses the basin-hopping algorithm from `scipy.optimize` to solve a global minimization problem, yielding the sequence of phase masks that minimizes the infidelity with respect to a target unitary. - [`jax_optimizer`](src/phaseshift/optimization/jax_optimizer.py): Uses `Jax` and `Optax` to perform gradient-based optimization of the phase masks with multiple restarts to minimize the infidelity or the geodesic distance [(Álvarez-Vizoso *et al.*)]( https://doi.org/10.48550/arXiv.2510.19397 ) with respect to a target unitary. This algorithm can run efficiently on CPU or GPU and is significantly faster than the SciPy-based implementation. --- **Note:** For more detailed descriptions and usage examples, see the documentation of the individual modules. ## Installation ### Install from PyPI (recommended) ```bash pip install phaseshift ``` ### Install from source You can install `PhaseShift` from source as follows: 1. Clone the repository ```bash git clone https://github.com/polyquantique/phaseshift.git cd phaseshift ``` 2. (Optional) Create and activate a virtual environment - Linux / macOS: ```bash python3 -m venv venv source venv/bin/activate ``` - Windows (Command Prompt): ```bash python -m venv venv venv\Scripts\activate ``` 3. Install the package and dependencies - Standard installation: ```bash pip install . ``` - Editable (developer) installation: ```bash pip install -e .[dev] ``` 4. (Optional) Run the tests ```bash pip install pytest pytest tests ``` ### Note For GPU support with JAX on Linux, install the appropriate CUDA-enabled version: ```bash pip install "jax[cuda]" -f https://storage.googleapis.com/jax-releases/jax_cuda_releases.html ``` ## Usage ### 1. `clements_interferometer` module This first example shows how to use the `clements_interferometer` module to decompose a random unitary matrix. ```python >>> from phaseshift import clements_interferometer as ci >>> from scipy.stats import unitary_group >>> import numpy as np ``` The function `clements_decomposition` performs the decomposition described in [Clements *et al.*, 2016](https://doi.org/10.1364/OPTICA.3.001460) on a random unitary matrix `U`. ```python >>> # Generate a random unitary matrix >>> U = unitary_group(dim = 8, seed = 137).rvs() >>> # Compute the Clements decomposition >>> decomposition = ci.clements_decomposition(U) ``` The output of this function is a `Decomposition` object, which has a `circuit` attribute. `Decomposition.circuit` is a list of `MachZehnder` objects that contain the parameters $\theta$ and $\phi$ of each unit cell in the mesh. ```python >>> # Extract the circuit from the decomposition >>> circuit = decomposition.circuit >>> # Print the parameters of the first unit cell in the circuit >>> print(circuit[0]) MachZehnder(theta=np.float64(0.7697802543915319), phi=np.float64(3.8400842306814207), target=(5, 6)) ``` The function `circuit_reconstruction` allows to compute the matrix that corresponds to a `Decomposition` object. This matrix can be compared to the original matrix. ```python >>> # Reconstruct the unitary matrix from the decomposition >>> reconstructed_matrix = ci.circuit_reconstruction(decomposition) >>> # Compare with the initial matrix >>> print(np.allclose(U, reconstructed_matrix)) True ``` ### 2. `fourier_interferometer` module This example shows how to use the `fourier_interferometer` module to decompose a random unitary matrix. ```python >>> from phaseshift import fourier_interferometer as fi >>> from scipy.stats import unitary_group >>> import numpy as np ``` The `compact_fourier_decomposition` function decomposes a random $N \times N$ unitary matrix into a sequence of $2N+5$ phase masks and $2N+4$ DFT layers. ```python >>> # Generate a random unitary matrix >>> U = unitary_group(dim = 8, seed = 137).rvs() >>> # Compute the Compact Fourier decomposition >>> decomposition = fi.compact_fourier_decomposition(U) ``` The output is a `FourierDecomp` object, which contains the `mask_sequence` attribute. `FourierDecomp.mask_sequence` stores the sequence of phase masks to be interleaved with DFT matrices. ```python >>> # Extract the mask sequence from the decomposition >>> mask_sequence = decomposition.mask_sequence >>> # Print the first mask in the sequence >>> print(mask_sequence[0].round(3)) [0.707-0.707j 0.707+0.707j 0.707-0.707j 0.707+0.707j 0.707-0.707j 0.707+0.707j 0.707-0.707j 0.707+0.707j] ``` The function `circuit_reconstruction` computes the matrix given by a `FourierDecomp` object by inserting a DFT matrix between each phase mask. The matrix can then be compared to the initial matrix. ```python >>> # Reconstruct the unitary matrix from the decomposition >>> reconstructed_matrix = fi.circuit_reconstruction(decomposition) >>> # Compare with the initial matrix >>> print(np.allclose(U, reconstructed_matrix)) True ``` ### 3. `optimization.jax_optimizer` module This example shows how to use the `jax_optimizer` module found in the `optimization` subpackage to find numerically a sequence of phase masks that approximate a given unitary matrix. ```python >>> from phaseshift.optimize import jax_optimizer as jo >>> from scipy.stats import unitary_group >>> import numpy as np ``` The function `jax_mask_optimizer` uses an Adam optimizer with multiple restarts to optimize a sequence of phase masks of a given length. In this case, we optimize 9 phase masks to ensure a complete parametrization of an $8 \times 8$ random unitary `U`. ```python >>> # Generate a random unitary matrix >>> U = unitary_group(dim = 8, seed = 137).rvs() >>> # Compute the phase masks that minimize infidelity >>> decomp, infidelity = jo.jax_mask_optimizer(U, length=9, steps=3500, restarts=50) ``` The function returns the decomposition, given as a `FourierDecomp` object, as well as the final infidelity obtained by the optimizer. ```python >>> # Print the final infidelity >>> print(infidelity) 2.220446049250313e-16 >>> # Print the type of decomp >>> print(type(decomp).__name__) FourierDecomp ``` The final matrix can be reconstructed using the `circuit_reconstruction` function from the `fourier_interferometer` module. ## Documentation The **LPLM algorithm** found in the [`lplm_interferometer`](src/phaseshift/lplm_interferometer.py) module was adapted from [López Pastor *et al.*, 2021](https://doi.org/10.1364/OE.432787) and uses a slightly different sequence of phase masks than the original paper. A comprehensive derivation of this new sequence can be found in the following document: - [Decomposition of Unitary Matrices Using Fourier Transforms and Phase Masks](https://github.com/polyquantique/phaseshift/raw/main/papers/LPLM_algorithm_derivation.pdf) ## Citing This Work If you find our research useful in your work, please cite it as ``` @article{girouard2025, author = {Vincent Girouard and Nicol\'{a}s Quesada}, journal = {J. Opt. Soc. Am. B}, number = {3}, pages = {A66--A73}, title = {Near-optimal decomposition of unitary matrices using phase masks and the discrete Fourier transform}, volume = {43}, year = {2026}, url = {https://opg.optica.org/josab/abstract.cfm?URI=josab-43-3-A66}, doi = {10.1364/JOSAB.577579} } ``` ## License This project is licensed under the Apache License 2.0. See the [LICENSE](https://github.com/polyquantique/phaseshift/blob/main/LICENSE) file for details. ## References - Girouard, Vincent, Nicolas, Quesada. "Near-optimal decomposition of unitary matrices using phase masks and the discrete Fourier transform". JOSA B Vol. 43, Issue 3, (2026): A66-A73 - Clements, William R., et al. "Optimal design for universal multiport interferometers." Optica 3.12 (2016): 1460-1465. - Bell, Bryn A., and Ian A. Walmsley. "Further compactifying linear optical unitaries." Apl Photonics 6.7 (2021). - López Pastor, Víctor, Jeff Lundeen, and Florian Marquardt. "Arbitrary optical wave evolution with Fourier transforms and phase masks." Optics Express 29.23 (2021): 38441-38450. - Saygin, M. Yu, et al. "Robust architecture for programmable universal unitaries." Physical review letters 124.1 (2020): 010501. - Pereira, Luciano, et al. "Minimum optical depth multiport interferometers for approximating arbitrary unitary operations and pure states." Physical Review A 111.6 (2025): 062603.
text/markdown
null
Vincent Girouard <vincent-2.girouard@polymtl.ca>, Nicolás Quesada <nicolas.quesada@polymtl.ca>
null
null
Apache License Version 2.0, January 2004 http://www.apache.org/licenses/ TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION 1. Definitions. "License" shall mean the terms and conditions for use, reproduction, and distribution as defined by Sections 1 through 9 of this document. "Licensor" shall mean the copyright owner or entity authorized by the copyright owner that is granting the License. "Legal Entity" shall mean the union of the acting entity and all other entities that control, are controlled by, or are under common control with that entity. For the purposes of this definition, "control" means (i) the power, direct or indirect, to cause the direction or management of such entity, whether by contract or otherwise, or (ii) ownership of fifty percent (50%) or more of the outstanding shares, or (iii) beneficial ownership of such entity. "You" (or "Your") shall mean an individual or Legal Entity exercising permissions granted by this License. "Source" form shall mean the preferred form for making modifications, including but not limited to software source code, documentation source, and configuration files. "Object" form shall mean any form resulting from mechanical transformation or translation of a Source form, including but not limited to compiled object code, generated documentation, and conversions to other media types. "Work" shall mean the work of authorship, whether in Source or Object form, made available under the License, as indicated by a copyright notice that is included in or attached to the work (an example is provided in the Appendix below). "Derivative Works" shall mean any work, whether in Source or Object form, that is based on (or derived from) the Work and for which the editorial revisions, annotations, elaborations, or other modifications represent, as a whole, an original work of authorship. For the purposes of this License, Derivative Works shall not include works that remain separable from, or merely link (or bind by name) to the interfaces of, the Work and Derivative Works thereof. "Contribution" shall mean any work of authorship, including the original version of the Work and any modifications or additions to that Work or Derivative Works thereof, that is intentionally submitted to Licensor for inclusion in the Work by the copyright owner or by an individual or Legal Entity authorized to submit on behalf of the copyright owner. For the purposes of this definition, "submitted" means any form of electronic, verbal, or written communication sent to the Licensor or its representatives, including but not limited to communication on electronic mailing lists, source code control systems, and issue tracking systems that are managed by, or on behalf of, the Licensor for the purpose of discussing and improving the Work, but excluding communication that is conspicuously marked or otherwise designated in writing by the copyright owner as "Not a Contribution." "Contributor" shall mean Licensor and any individual or Legal Entity on behalf of whom a Contribution has been received by Licensor and subsequently incorporated within the Work. 2. Grant of Copyright License. Subject to the terms and conditions of this License, each Contributor hereby grants to You a perpetual, worldwide, non-exclusive, no-charge, royalty-free, irrevocable copyright license to reproduce, prepare Derivative Works of, publicly display, publicly perform, sublicense, and distribute the Work and such Derivative Works in Source or Object form. 3. Grant of Patent License. Subject to the terms and conditions of this License, each Contributor hereby grants to You a perpetual, worldwide, non-exclusive, no-charge, royalty-free, irrevocable (except as stated in this section) patent license to make, have made, use, offer to sell, sell, import, and otherwise transfer the Work, where such license applies only to those patent claims licensable by such Contributor that are necessarily infringed by their Contribution(s) alone or by combination of their Contribution(s) with the Work to which such Contribution(s) was submitted. If You institute patent litigation against any entity (including a cross-claim or counterclaim in a lawsuit) alleging that the Work or a Contribution incorporated within the Work constitutes direct or contributory patent infringement, then any patent licenses granted to You under this License for that Work shall terminate as of the date such litigation is filed. 4. Redistribution. You may reproduce and distribute copies of the Work or Derivative Works thereof in any medium, with or without modifications, and in Source or Object form, provided that You meet the following conditions: (a) You must give any other recipients of the Work or Derivative Works a copy of this License; and (b) You must cause any modified files to carry prominent notices stating that You changed the files; and (c) You must retain, in the Source form of any Derivative Works that You distribute, all copyright, patent, trademark, and attribution notices from the Source form of the Work, excluding those notices that do not pertain to any part of the Derivative Works; and (d) If the Work includes a "NOTICE" text file as part of its distribution, then any Derivative Works that You distribute must include a readable copy of the attribution notices contained within such NOTICE file, excluding those notices that do not pertain to any part of the Derivative Works, in at least one of the following places: within a NOTICE text file distributed as part of the Derivative Works; within the Source form or documentation, if provided along with the Derivative Works; or, within a display generated by the Derivative Works, if and wherever such third-party notices normally appear. The contents of the NOTICE file are for informational purposes only and do not modify the License. You may add Your own attribution notices within Derivative Works that You distribute, alongside or as an addendum to the NOTICE text from the Work, provided that such additional attribution notices cannot be construed as modifying the License. You may add Your own copyright statement to Your modifications and may provide additional or different license terms and conditions for use, reproduction, or distribution of Your modifications, or for any such Derivative Works as a whole, provided Your use, reproduction, and distribution of the Work otherwise complies with the conditions stated in this License. 5. Submission of Contributions. Unless You explicitly state otherwise, any Contribution intentionally submitted for inclusion in the Work by You to the Licensor shall be under the terms and conditions of this License, without any additional terms or conditions. Notwithstanding the above, nothing herein shall supersede or modify the terms of any separate license agreement you may have executed with Licensor regarding such Contributions. 6. Trademarks. This License does not grant permission to use the trade names, trademarks, service marks, or product names of the Licensor, except as required for reasonable and customary use in describing the origin of the Work and reproducing the content of the NOTICE file. 7. Disclaimer of Warranty. Unless required by applicable law or agreed to in writing, Licensor provides the Work (and each Contributor provides its Contributions) on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied, including, without limitation, any warranties or conditions of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A PARTICULAR PURPOSE. You are solely responsible for determining the appropriateness of using or redistributing the Work and assume any risks associated with Your exercise of permissions under this License. 8. Limitation of Liability. In no event and under no legal theory, whether in tort (including negligence), contract, or otherwise, unless required by applicable law (such as deliberate and grossly negligent acts) or agreed to in writing, shall any Contributor be liable to You for damages, including any direct, indirect, special, incidental, or consequential damages of any character arising as a result of this License or out of the use or inability to use the Work (including but not limited to damages for loss of goodwill, work stoppage, computer failure or malfunction, or any and all other commercial damages or losses), even if such Contributor has been advised of the possibility of such damages. 9. Accepting Warranty or Additional Liability. While redistributing the Work or Derivative Works thereof, You may choose to offer, and charge a fee for, acceptance of support, warranty, indemnity, or other liability obligations and/or rights consistent with this License. However, in accepting such obligations, You may act only on Your own behalf and on Your sole responsibility, not on behalf of any other Contributor, and only if You agree to indemnify, defend, and hold each Contributor harmless for any liability incurred by, or claims asserted against, such Contributor by reason of your accepting any such warranty or additional liability. END OF TERMS AND CONDITIONS APPENDIX: How to apply the Apache License to your work. To apply the Apache License to your work, attach the following boilerplate notice, with the fields enclosed by brackets "[]" replaced with your own identifying information. (Don't include the brackets!) The text should be enclosed in the appropriate comment syntax for the file format. We also recommend that a file or class name and description of purpose be included on the same "printed page" as the copyright notice for easier identification within third-party archives. Copyright [yyyy] [name of copyright owner] Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.
null
[]
[]
null
null
>=3.8
[]
[]
[]
[ "numpy", "scipy", "jax>=0.7.0", "optax>=0.2.5", "pytest>=7.0; extra == \"dev\"", "pytest-cov>=4.0.0; extra == \"dev\"", "black>=23.0; extra == \"dev\"", "isort; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://github.com/polyquantique/phaseshift" ]
twine/6.2.0 CPython/3.11.9
2026-02-20T21:37:06.665471
phaseshift-1.0.0.tar.gz
48,729
55/c6/3378a84b1e8139096abfdaccfaf360095fbe0267b6fa2694a78f122c1688/phaseshift-1.0.0.tar.gz
source
sdist
null
false
4e385ff1b47135e6c8c7a0580952f4c6
4024e063f7faf91eb4ee9663e1c48daae1bf052f23254421f955bdd04b695f3e
55c63378a84b1e8139096abfdaccfaf360095fbe0267b6fa2694a78f122c1688
null
[ "LICENSE" ]
220
2.4
agilix-api-fr8train
2.1.2
A Python SDK for integrating with the Agilix Buzz API
# Agilix API ### By fr8train-sv A Python SDK for integrating with Agilix Buzz API. Install with: ```bash pip install agilix-api-fr8train ``` Upgrade this specific library by running: ```bash pip install --upgrade agilix-api-fr8train ``` ## Important Commands The following list of commands are important for project maintenance. ## Building the Library To build the library, ensure that the necessary build tools are installed in your environment. This can be done by installing `setuptools`, `build`, and `wheel`: ```bash pip install build twine setuptools wheel ``` `twine` can also be installed here while we're installing shit since we'll need it later. **REMEMBER**: INCREMENT YOUR VERSION NUMBER IN THE TOML BEFORE BUILDING. **REMEMBER**: REMOVE THE /DIST DIRECTORY BEFORE BUILDING FOR A CLEAN BUILD Now, you can create the distribution files (source distribution and wheel) using the following command: ```bash python3 -m build ``` This will generate builds in the `dist/` directory. ## Deploying to PyPI Ensure you have a valid PyPI account and credentials added to your `.pypirc` file or provide them during the publish process. Then, upload your package to PyPI with: ```bash python3 -m twine upload dist/* ``` Follow any prompts from `twine` to successfully upload your package. Once deployed, your package will be available on PyPI. # Usage ## Initialization To start using the API SDK, you will need to initialize an API object: ```python from agilix_api_fr8train.api import Api api = Api() ``` This will reach out into your project for a .env file for credentials to establish a connection to the Agilix API Gateway. Structure your .env file like the example below and maintain it at the root of your project directory. ```dotenv AGILIX_BASE_URL= AGILIX_DOMAIN= AGILIX_USERNAME= AGILIX_PASSWORD= AGILIX_HOME_DOMAIN_ID= ```
text/markdown
null
Tyler Collette <tyler.collette@gmail.com>
null
null
null
agilix, api, sdk, buzz, integration
[ "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.12", "Operating System :: OS Independent" ]
[]
null
null
null
[]
[]
[]
[ "requests", "dotenv", "python-dotenv", "pendulum" ]
[]
[]
[]
[ "Documentation, https://github.com/fr8train-sv/agilix-api/wiki", "Source, https://github.com/fr8train-sv/agilix-api", "Changelog, https://github.com/fr8train-sv/agilix-api/releases" ]
twine/6.2.0 CPython/3.12.3
2026-02-20T21:36:06.437077
agilix_api_fr8train-2.1.2.tar.gz
16,181
9d/6c/6ab218fd1dee9d51d470c2d4f887f65aa73a3293f58d7a3163b8e05dcdfd/agilix_api_fr8train-2.1.2.tar.gz
source
sdist
null
false
f659cf9fc847b816145a5a8deaf3b753
8d2232e3d4c8d29f854b50af24bef1e2cdef295b9ae01f3a3d93ef3f164651e9
9d6c6ab218fd1dee9d51d470c2d4f887f65aa73a3293f58d7a3163b8e05dcdfd
null
[ "LICENSE" ]
214
2.4
lshrs
0.1.1b2
Redis-backed Locality Sensitive Hashing toolkit for fast approximate nearest neighbor search
# LSHRS [![CI](https://github.com/mxngjxa/lshrs/actions/workflows/ci.yml/badge.svg)](https://github.com/mxngjxa/lshrs/actions/workflows/ci.yml) [![Publish to PyPI](https://github.com/mxngjxa/lshrs/actions/workflows/cd.yml/badge.svg)](https://github.com/mxngjxa/lshrs/actions/workflows/cd.yml) [![PyPI version](https://img.shields.io/pypi/v/lshrs.svg)](https://pypi.org/project/lshrs/) [![Python](https://img.shields.io/badge/python-3.10%2B-blue.svg)](https://www.python.org/downloads/) [![MIT License](https://img.shields.io/badge/license-MIT-green.svg)](LICENSE) [![Ruff](https://img.shields.io/endpoint?url=https://raw.githubusercontent.com/astral-sh/ruff/main/assets/badge/v2.json)](https://github.com/astral-sh/ruff) Redis-backed locality-sensitive hashing toolkit that stores bucket membership in Redis while keeping the heavy vector payloads in your primary datastore. <div align="center"> <img src="docs/lshrs-logo.svg" alt="logo"></img> </div> ## Table of Contents - [Overview](#overview) - [Architecture Snapshot](#architecture-snapshot) - [Key Features](#key-features) - [Installation](#installation) - [Quick Start](#quick-start) - [Ingestion Pipelines](#ingestion-pipelines) - [Querying Modes](#querying-modes) - [Persistence & Lifecycle](#persistence--lifecycle) - [Performance & Scaling Guidelines](#performance--scaling-guidelines) - [Troubleshooting](#troubleshooting) - [API Surface Summary](#api-surface-summary) - [Development & Testing](#development--testing) - [License](#license) ## Overview [`LSHRS`](lshrs/core/main.py:53) orchestrates the full locality-sensitive hashing (LSH) workflow: 1. Hash incoming vectors into stable banded signatures via random projections. 2. Store only bucket membership in Redis for low-latency candidate enumeration. 3. Optionally rerank candidates using cosine similarity with vectors fetched from your system of record. The out-of-the-box configuration chooses bands/rows automatically, pipelines Redis operations, and exposes hooks for streaming data ingestion, persistence, and operational maintenance. ## Architecture Snapshot | Concern | Component | Description | | --- | --- | --- | | Hashing | [`LSHHasher`](lshrs/hash/lsh.py:20) | Generates banded random-projection signatures. | | Storage | [`RedisStorage`](lshrs/storage/redis.py:40) | Persists bucket membership using Redis sets and pipelines for batch writes. | | Ingestion | [`LSHRS.create_signatures()`](lshrs/core/main.py:267) | Streams vectors from PostgreSQL or Parquet via pluggable loaders. | | Reranking | [`top_k_cosine()`](lshrs/utils/similarity.py:94) | Computes cosine similarity for candidate reranking. | | Configuration | [`get_optimal_config()`](lshrs/utils/br.py:326) | Picks band/row counts that match a target similarity threshold. | ## Key Features - **Redis-native buckets**: Uses Redis sets for O(1) membership updates and pipelined batch ingestion. - **Progressive indexing**: Stream vectors from PostgreSQL ([`iter_postgres_vectors()`](lshrs/io/postgres.py:16)) or Parquet ([`iter_parquet_vectors()`](lshrs/io/parquet.py:46)) without exhausting memory. - **Dual retrieval modes**: Choose fast top-k collision lookups or cosine-reranked top-p filtering through [`LSHRS.query()`](lshrs/core/main.py:486). - **Persistable hashing state**: Save and reload projection matrices with [`LSHRS.save_to_disk()`](lshrs/core/main.py:830) and [`LSHRS.load_from_disk()`](lshrs/core/main.py:881). - **Operational safety**: Snapshot configuration with [`LSHRS.stats()`](lshrs/core/main.py:782), clear indices via [`LSHRS.clear()`](lshrs/core/main.py:755), and surgically delete members using [`LSHRS.delete()`](lshrs/core/main.py:710). ## Installation ### PyPI ```bash pip install lshrs ``` Or, if installing for a postgres database: ```bash pip install 'lshrs[postgres]' ``` Or with Parquet ingestion support: ```bash pip install 'lshrs[parquet]' ``` ### From source checkout ```bash git clone https://github.com/mxngjxa/lshrs.git cd lshrs uv sync -e ".[dev]" ``` > [!NOTE] > The project requires Python >= 3.10 as defined in [`pyproject.toml`](pyproject.toml). ### Optional extras - PostgreSQL streaming requires [`psycopg`](https://www.psycopg.org/). Install with `pip install 'lshrs[postgres]'`. - Parquet ingestion requires [`pyarrow`](https://arrow.apache.org/). Install with `pip install 'lshrs[parquet]'`. ## Quick Start ```python import numpy as np from lshrs import LSHRS def fetch_vectors(indices: list[int]) -> np.ndarray: # Replace with your vector store retrieval (PostgreSQL, disk, object store, etc.) embeddings = np.load("vectors.npy") return embeddings[indices] lsh = LSHRS( dim=768, num_perm=256, redis_host="localhost", redis_prefix="demo", vector_fetch_fn=fetch_vectors, ) # Stream index construction from PostgreSQL lsh.create_signatures( format="postgres", dsn="postgresql://user:pass@localhost/db", table="documents", index_column="doc_id", vector_column="embedding", ) # Insert an ad-hoc document lsh.ingest(42, np.random.randn(768).astype(np.float32)) # Retrieve candidates query = np.random.randn(768).astype(np.float32) top10 = lsh.get_top_k(query, topk=10) reranked = lsh.get_above_p(query, p=0.2) ``` The code above exercises [`LSHRS.create_signatures()`](lshrs/core/main.py:267), [`LSHRS.ingest()`](lshrs/core/main.py:340), [`LSHRS.get_top_k()`](lshrs/core/main.py:626), and [`LSHRS.get_above_p()`](lshrs/core/main.py:661). ## Ingestion Pipelines ### Streaming from PostgreSQL [`iter_postgres_vectors()`](lshrs/io/postgres.py:16) yields `(indices, vectors)` batches using server-side cursors: ```python lsh.create_signatures( format="postgres", dsn="postgresql://reader:secret@analytics.db/search", table="embeddings", index_column="item_id", vector_column="embedding", batch_size=5_000, where_clause="updated_at >= NOW() - INTERVAL '1 day'", ) ``` > [!TIP] > Provide a custom `connection_factory` if you need pooled connections or TLS configuration. ### Streaming from Parquet [`iter_parquet_vectors()`](lshrs/io/parquet.py:46) supports memory-friendly batch loads from Parquet files: ```python for ids, batch in iter_parquet_vectors( "captures/2024-01-embeddings.parquet", index_column="document_id", vector_column="embedding", batch_size=8_192, ): lsh.index(ids, batch) ``` > [!IMPORTANT] > Install `pyarrow` prior to using the Parquet loader; otherwise [`iter_parquet_vectors()`](lshrs/io/parquet.py:46) raises `ImportError`. ### Manual or Buffered Ingestion - [`LSHRS.index()`](lshrs/core/main.py:399) ingests vector batches you already hold in memory. - [`LSHRS.ingest()`](lshrs/core/main.py:340) is ideal for realtime single-document updates. - Under the hood, [`RedisStorage.batch_add()`](lshrs/storage/redis.py:340) leverages Redis pipelines for throughput. ## Querying Modes [`LSHRS.query()`](lshrs/core/main.py:486) provides two complementary retrieval patterns: | Mode | When to use | Result | | --- | --- | --- | | **Top-k** (`top_p=None`) | Latency-critical scenarios that only require coarse candidates. | Returns `List[int]` ordered by band collisions. | | **Top-p** (`top_p=0.0–1.0`) | Precision-sensitive flows that can rerank using original vectors. | Returns `List[Tuple[int,float]]` of `(index, cosine_similarity)` pairs. | > [!CAUTION] > Reranking requires configuring `vector_fetch_fn` when instantiating [`LSHRS`](lshrs/core/main.py:53); otherwise top-p queries raise `RuntimeError`. Supporting helpers: - [`LSHRS.get_top_k()`](lshrs/core/main.py:626) wraps `query` for pure top-k retrieval. - [`LSHRS.get_above_p()`](lshrs/core/main.py:661) wraps `query` with a similarity-mass cutoff. - Cosine scoring is provided by [`cosine_similarity()`](lshrs/utils/similarity.py:25) and [`top_k_cosine()`](lshrs/utils/similarity.py:94). ## Persistence & Lifecycle | Operation | Purpose | Reference | | --- | --- | --- | | Snapshot configuration | Inspect runtime parameters and Redis namespace. | [`LSHRS.stats()`](lshrs/core/main.py:782) | | Flush & clear | Remove all Redis buckets for the configured prefix. | [`LSHRS.clear()`](lshrs/core/main.py:755) | | Hard delete members | Remove specific indices across all buckets. | [`LSHRS.delete()`](lshrs/core/main.py:710) | | Persist projections | Save configuration and projection matrices to disk. | [`LSHRS.save_to_disk()`](lshrs/core/main.py:830) | | Restore projections | Rebuild an instance using saved matrices. | [`LSHRS.load_from_disk()`](lshrs/core/main.py:881) | > [!WARNING] > [`LSHRS.clear()`](lshrs/core/main.py:755) is irreversible—every key with the configured prefix is deleted. Back up state with [`LSHRS.save_to_disk()`](lshrs/core/main.py:830) beforehand if you need to rebuild. ## Performance & Scaling Guidelines - **Choose sensible hash parameters**: [`get_optimal_config()`](lshrs/utils/br.py:326) finds bands/rows that approximate your target similarity threshold. Inspect S-curve behavior with [`compute_collision_probability()`](lshrs/utils/br.py:119). - **Normalize inputs**: Pre-normalize vectors or rely on [`l2_norm()`](lshrs/utils/norm.py:4) for consistent cosine scores. - **Batch ingestion**: When indexing large volumes, route operations through [`LSHRS.index()`](lshrs/core/main.py:399) to let [`RedisStorage.batch_add()`](lshrs/storage/redis.py:340) coalesce writes. - **Monitor bucket sizes**: Large buckets indicate low selectivity. Adjust `num_perm`, `num_bands`, or the similarity threshold to trade precision vs. recall. - **Pipeline warmup**: Flush outstanding operations with [`LSHRS._flush_buffer()`](lshrs/core/main.py:1177) (indirectly called) before measuring latency or persisting state. ## Troubleshooting | Symptom | Likely Cause | Resolution | | --- | --- | --- | | `ImportError: psycopg is required` | PostgreSQL loader invoked without optional dependency. | Install `psycopg[binary]` or avoid `format="postgres"`. | | `ValueError: Vectors must have shape (n, dim)` | Supplied batch dimension mismatched the configured `dim`. | Ensure all vectors match the `dim` passed to [`LSHRS.__init__()`](lshrs/core/main.py:149). | | `ValueError: Cannot normalize zero vector` | Zero-length vectors were passed to cosine scoring utilities. | Filter zero vectors before reranking or normalize upstream. | | Empty search results | Buckets never flushed to Redis. | Call [`LSHRS.index()`](lshrs/core/main.py:399) (auto flushes) or explicitly invoke [`LSHRS._flush_buffer()`](lshrs/core/main.py:1177) before querying. | | Extremely large buckets | Similarity threshold too low / insufficient hash bits. | Increase `num_perm` or tweak target threshold via [`get_optimal_config()`](lshrs/utils/br.py:326). | > [!TIP] > Use Redis `SCAN` commands (e.g., `SCAN 0 MATCH lsh:*`) to inspect bucket distribution during tuning. ## API Surface Summary | Area | Description | Primary Entry Point | | --- | --- | --- | | Ingestion orchestration | Bulk streaming with source-aware loaders. | [`LSHRS.create_signatures()`](lshrs/core/main.py:267) | | Batch ingestion | Hash and store vectors already in memory. | [`LSHRS.index()`](lshrs/core/main.py:399) | | Single ingestion | Add or update one vector id on the fly. | [`LSHRS.ingest()`](lshrs/core/main.py:340) | | Candidate enumeration | General-purpose search with optional reranking. | [`LSHRS.query()`](lshrs/core/main.py:486) | | Hash persistence | Save and restore LSH projection matrices. | [`LSHRS.save_to_disk()`](lshrs/core/main.py:830) / [`LSHRS.load_from_disk()`](lshrs/core/main.py:881) | | Redis maintenance | Prefix-aware key deletion and batch removal. | [`RedisStorage.clear()`](lshrs/storage/redis.py:582) / [`RedisStorage.remove_indices()`](lshrs/storage/redis.py:411) | | Probability utilities | Analyze band/row trade-offs and false rates. | [`compute_collision_probability()`](lshrs/utils/br.py:119) / [`compute_false_rates()`](lshrs/utils/br.py:161) | ## Development & Testing 1. Clone and install development dependencies: ```bash git clone https://github.com/mxngjxa/lshrs.git cd lshrs uv sync --dev ``` 2. Run the test suite: ```bash uv run pytest ``` 3. Lint and format check: ```bash uv run ruff check . uv run ruff format --check . ``` > [!NOTE] > Example snippets in this README are intended to be run under Python >= 3.10 with NumPy >= 1.24 and Redis >= 7 as specified in [`pyproject.toml`](pyproject.toml). ## License Licensed under the terms of [`LICENSE`](LICENSE).
text/markdown
null
Mingjia Guan <mxngjxa@gmail.com>
null
null
null
approximate-nearest-neighbor, locality-sensitive-hashing, lsh, redis, similarity-search, vector-search
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Science/Research", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Database", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Typing :: Typed" ]
[]
null
null
>=3.10
[]
[]
[]
[ "numpy>=1.24", "redis>=7.0.1", "scipy>=1.14.1", "pyarrow>=14.0; extra == \"parquet\"", "psycopg>=3.2.12; extra == \"postgres\"" ]
[]
[]
[]
[ "Homepage, https://github.com/mxngjxa/lshrs", "Repository, https://github.com/mxngjxa/lshrs", "Documentation, https://github.com/mxngjxa/lshrs/blob/main/docs/docs.md", "Issues, https://github.com/mxngjxa/lshrs/issues", "Changelog, https://github.com/mxngjxa/lshrs/releases" ]
uv/0.10.4 {"installer":{"name":"uv","version":"0.10.4","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
2026-02-20T21:36:02.245153
lshrs-0.1.1b2.tar.gz
135,837
29/1a/3805b79fb71f695541e410901f249063e1edf410bd751de3d9a2ed07fa48/lshrs-0.1.1b2.tar.gz
source
sdist
null
false
641d3dceafd7ba4d928479f1cfd50d05
62536f161d4e2939ff60908067529a30efea18420c25ff240fd7f871e83e9854
291a3805b79fb71f695541e410901f249063e1edf410bd751de3d9a2ed07fa48
MIT
[ "LICENSE" ]
196
2.4
dawgie
2.1.0rc3
Data and Algorithm Work-flow Generation, Introspection, and Execution (DAWGIE)
## DAWGIE ### Data and Algorithm Work-flow Generation, Introspection, and Execution The DAWGIE software accomplishes: 1. Data anonymity is required by the framework because the implementation DAWGIE is independent of the Algorithm Engine element of the framework. The framework allows DAWGIE to identify the author of the data through naming and version even though DAWGIE as no knowledge of the data itself. Because the language of the framework and further implementations is Python, DAWGIE forces a further requirement on the data that it be "pickle-able". The additional requirement was accepted by the framework and pushed to the Algorithm Engine element. DAWGIE uses a lazy load implementation -- loaded only when an Algorithm Engine sub-element requests data from a specific author -- to remove routing and, thus, requiring knowledge about the data itself. 1. Data persistence has two independent implementations. In the case of small and simple data author relationships, the Python shelve module is used for persistence. As the number of unique authors grows, the relations grow faster (non-linear) and requires a more sophisticated persistence tool. In the latter case, postgresql is used for persistence. Both implementations are the design making them quickly interchangeable. Neither actually store the data blobs themselves. They -- shelve and postgresql -- only store the author's unique ID -- name and version -- and then reference the blob on disk. Using the relational DB proprieties of shelve (Python dictionaries) and postgresql (relational DB) any request for data from a specific author can quickly be resolved. Creating unique names for all of the data blobs is achieved using {md5 hash}_{sha1 hash}. These names also allow the persistence to recognize if the data is already known. 1. For pipeline management, DAWGIE implements a worker farm, a work flow Abstract Syntax Tree (AST), and signaling from data persistence to execute the tasks within the framework Algorithm Engine element. The data persistence implementation signals the pipeline management element when new data has been authored. The manager searches the AST for all tasks that depend on the author and schedules all tasks that depend on the author starting with the earliest dependent. The new data signals can be generated at the end any task. When a task moves from being scheduled to executing, the foreman of the worker farm passes the task to a waiting worker on a compute node. The worker then loads the data via data persistence for the task and begins executing the task. Upon completion of the task, the worker saves the data via data persistence and notifies the foreman it is ready for another task. In this fashion, DAWGIE walks the minimum and complete nodes in the AST that depend on any new data that has been generated. The pipeline management offers periodic tasks as well that treat a temporal event as new data. ### Organization DAWGIE is configured and controlled by ENV variables and command line arguments. The command line arguments override the ENV variables. While this section cover nearly all of the command line options, please use the --help switch to see all of the available options. #### Access to Running DAWGIE <dl> <dt>DAWGIE_DB_PORT --context-db-port</dt> <dd>The database access port. See the specific database implemntation being used for detailed definition of this parameter.</dd> <dt>DAWGIE_FARM_PORT --context-farm-port</dt> <dd>The access port that workers in the farm will use to communicate with the DAWGIE server.</dd> <dt>DAWGIE_FE_PORT --port</dt> <dd>The web display port. All subsequent ports are computed from this one.</dd> <dt>DAWGIE_LOG_PORT --context-log-port</dt> <dd>Port number for distributed workers to log messages through the DAWGIE server.</dd> </dl> #### Algorithm Engine (AE) <dl> <dt>DAWGIE_AE_BASE_PATH --context-ae-dir</dt> <dd>The complete path to the AE source code. It is the first directory to start walking down and checking all of the sub-directories and including those that are packages that implement the necessary factories to be identified as AE packages.</dd> <dt>DAAWGIE_AE_BASE_PACKAGE --context-ae-pkg</dt> <dd>Because the AE code may be intermixed with non-AE code and, therefore, may be a subset of code, need to know the packaage prefix.</dd> </dl> ** Example ** If all of the Python code starts in foo and the AE code starts in foo/bar/ae. Then `DAWGIE_AE_BASE_PATH` should be 'foo/bar/ae' and `DAWGIE_AE_BASE_PACKAGE` should be 'foo.bar.ae'. #### Data <dl> <dt>DAWGIE_DATA_DBSTOR --context-data-dbs</dt> <dd>The location for DAWGIE to store the data generated by the AE known as StateVectors. This area should be vast enough to hold all of the data genreated by the AE over all time.</dd> <dt>DAWGIE_DATA_LOGDIR --context-data-log</dt> <dd>The location for DAWGIE to write its log files.</dd> <dt>DAWGIE_DATA_STAGED --context-data-stg</dt> <dd>The location for DAWGIE to store temporary data for the AE. It should be sized to fit the expected AE use and DAWGIE will clean out the staging area. However, when there are hiccups in the </dd> </dl> #### Database DAWGIE supports two styles of databases. It supports Python shelve for tiny applications and then POSTGRES for a much larger and scalable system. ##### Postgresql <dl> <dt>DAWGIE_DB_HOST --context-db-host</dt> <dd>The IP hostname of the POSTGRESQL server.</dd> <dt>DAWGIE_DB_IMPL --context-db-impl</dt> <dd>Must be 'post'</dd> <dt>DAWGIE_DB_NAME --context-db-name</dt> <dd>The name of the database to use.</dd> <dt>DAWGIE_DB_PATH --context-db-path</dt> <dd>THe username:password for the database named with DAWGIE_DB_NAME.</dd> <dt>DAWGIE_DB_PORT --context-db-port</dt> <dd>The IP port number of the POSTGRESQL server. When DAWGIE_DB_IMPL is 'post', this value defaults to 5432 because POSTGRESQL is independent of DAWGIE.</dd> </dl> ##### Shelve <dl> <dt>DAWGIE_DB_HOST --context-db-host</dt> <dd>The IP hostname of the machine running DAWGIE.</dd> <dt>DAWGIE_DB_IMPL --context-db-impl</dt> <dd>Must be 'shelve'</dd> <dt>DAWGIE_DB_NAME --context-db-name</dt> <dd>The name of the database to use.</dd> <dt>DAWGIE_DB_PATH --context-db-path</dt> <dd>The directory path on DAWGIE_DB_HOST to write the shelve files.</dd> <dt>DAWGIE_DB_PORT --context-db-port</dt> <dd>The IP port number of the DAWGIE DB interface. In is automatically computed from the general port number where DAWGIE is being served (see --port)</dd> </dl> ##### Tools <dl> <dt>DAWGIE_DB_POST2SHELVE_PREFIX</dt> <dd>Used when converting POSTGRESQL to shelve for development of new AE modules.</dd> <dt>DAWGIE_DB_ROTATE_PATH --context-db-rotate-path</dt> <dd>Allows the data to backed up with every new run ID</dd> <dt>DAWGIE_DB_COPY_PATH --context-db-copy-path</dt> <dd>Temporary working space for database work.</dd> <dt>DAWGIE_DB_ROTATES --context-db-rotate</dt> <dd>The number of database backups to preserve.</dd> </dl> #### Souce Code The source code is then organized by language: - Bash : utilities for simpler access to the Python - Python : implementation of DAWGIE The Python has view key packages: - dawgie.db : the database interface - dawgie.de : the display engine that allows user requests to render state vectors to meaningful images - dawgie.fe : the [front-end](http://mentor.jpl.nasa.gov:8080) that we see and interact with - dawgie.pl : the actual pipeline code that exercises the algorithm engine - dawgie.tools : a tool box used by the pipeline and adminsitrators (mostly) ### Documentation [Fundamental Brochure](https://github.jpl.nasa.gov/pages/niessner/DAWGIE/Notebook/Fundamentals-Brochure.slides.html) is a sales brochure used for SOYA 2018 contest. [Fundamental Developer Overiew](https://github.jpl.nasa.gov/pages/niessner/DAWGIE/Notebook/Fundamentals-Developer-Overview.slides.html) is a mix of sales and development. It frames the problem and the solution provided. It then proceeds to a high level description of how to use the tool using Gamma et al *Design Patterns*. Armed with the patterns being used, a developer should be able to move to the HOW TO slides connecting the minutiae in those slides with the highest view in these. [Fundamental Magic](https://github.jpl.nasa.gov/pages/niessner/DAWGIE/Notebook/Fundamentals-Magic.slides.html) is a manager level explanation of what the pipeline does and how it can help development. [Fundamental How To](https://github.jpl.nasa.gov/pages/niessner/DAWGIE/Notebook/Fundamentals-HOWTO.slides.html) is a beginner course on working with DAWGIE. [Fundamental Administration](https://github.jpl.nasa.gov/pages/niessner/DAWGIE/Notebook/Fundamentals-Admin.slides.html) is a starter on how to adminsiter DAWGIE. ### Installation 1. `python3 Python/setup.py build install` 1. `bash install.sh` ### Use
text/markdown
Al Niessner
Al.Niessner@jpl.nasa.gov
null
null
see LICENSE file for details
adaptive pipeline
[ "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.12", "Operating System :: OS Independent", "License :: Free To Use But Restricted", "Development Status :: 5 - Production/Stable" ]
[]
https://github.com/al-niessner/DAWGIE
null
>=3.10
[]
[]
[]
[ "bokeh>=1.2", "boto3>=1.7.80", "cryptography>=2.1.4", "Deprecated", "GitPython>=2.1.11", "matplotlib>=2.1.1", "progressbar2", "psycopg>=3.2.12", "psycopg-binary>=3.2.12", "pydot", "pyparsing>=2.4.7", "pyOpenSSL>=19.1.0", "python-gnupg==0.4.9", "service_identity", "requests>=2.20.0", "transitions", "twisted>=24.3.0" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:35:48.868447
dawgie-2.1.0rc3-py3-none-any.whl
844,160
49/b7/7339a47c2be0fc768d9c916420f1ae508ea5bafa1689975fae50e3bb0952/dawgie-2.1.0rc3-py3-none-any.whl
py3
bdist_wheel
null
false
766a7aec95de5396918402c9c509edfb
362d54a09ae2e8eba10a08edf90873236c054fef0e89751856b87d7ca892e7aa
49b77339a47c2be0fc768d9c916420f1ae508ea5bafa1689975fae50e3bb0952
null
[]
75
2.4
ndevio
0.9.0
Read, write, and manage images in napari
# ndevio [![License BSD-3](https://img.shields.io/pypi/l/ndevio.svg?color=green)](https://github.com/ndev-kit/ndevio/raw/main/LICENSE) [![PyPI](https://img.shields.io/pypi/v/ndevio.svg?color=green)](https://pypi.org/project/ndevio) [![Python Version](https://img.shields.io/pypi/pyversions/ndevio.svg?color=green)](https://python.org) [![tests](https://github.com/ndev-kit/ndevio/workflows/tests/badge.svg)](https://github.com/ndev-kit/ndevio/actions) [![codecov](https://codecov.io/gh/ndev-kit/ndevio/branch/main/graph/badge.svg)](https://codecov.io/gh/ndev-kit/ndevio) [![napari hub](https://img.shields.io/endpoint?url=https://api.napari-hub.org/shields/ndevio)](https://napari-hub.org/plugins/ndevio) [![npe2](https://img.shields.io/badge/plugin-npe2-blue?link=https://napari.org/stable/plugins/index.html)](https://napari.org/stable/plugins/index.html) [![Copier](https://img.shields.io/endpoint?url=https://raw.githubusercontent.com/copier-org/copier/master/img/badge/badge-grayscale-inverted-border-purple.json)](https://github.com/copier-org/copier) **A generalized image format reader for napari built on top of [bioio]** `ndevio` provides flexible, metadata-aware image I/O for images in napari. Originally developed as part of napari-ndev (as a spiritual successor to [napari-aicsimageio]), `ndevio` has since been separated into its own plugin as part of the [ndev-kit] and is intended to be a feature-rich, metadata-aware image reader for a variety of formats in napari, with a focus on microscopy images. ---------------------------------- ## Features - **Extensive format support** via [bioio] and its plugin system — read OME-TIFF, OME-Zarr, common image and movie formats, proprietary formats (CZI, LIF, ND2), and many more (with bioformats)! - **Multi-scene handling** — interactive widget for selecting between scenes/positions in multi-scene files - **Thorough metadata extraction** — extract and apply scale, units, axis labels, metadata (inc. OME) to napari layers - **Remote file support** — compatible Bioio readers,such as [bioio-ome-zarr], can read from remote filesystems (HTTP, S3, etc.) with dask-backed loading - **Native multiscale support** — automatically read and display multiscale images when supported by the reader. For best experience, turn on the asynchronous rendering experimental setting in napari. - **Configurable behavior** via [ndev-settings] — customize reader priority, multi-scene handling, and more - **Smart plugin installation** — automatic suggestions to install missing bioio reader plugins - **Programmatic API** — `nImage` class for napari-ready metadata extraction - **Batch utilities** — legacy widget for batch concatenation (with [nbatch]) and metadata management, with features being superseded by [napari-metadata] - **Sample data** — demonstrates ndevio metadata handling and capabilities ![napari viewer showing scene selection widget and napari-metadata widget displaying the metadata provided by ndevio a .czi file](https://github.com/ndev-kit/ndevio/blob/main/resources/ndevio-scene-and-metadata.png?raw=true) ## Installation You can install `ndevio` from [PyPI] or via the napari plugin manager: ```bash pip install ndevio ``` If you would like to try out ndevio, you can run napari in a temporary environment with [uv]: ```bash uvx --with ndevio -p 3.13 "napari[all]" ``` To contibute to ndevio or experiment with the latest features, see [Contributing.md](CONTRIBUTING.md) for development setup instructions. Conda-forge availability is coming soon! ### Additional Image Format Support **ndevio** uses [bioio](https://github.com/bioio-devs/bioio) for flexible image reading. Basic formats (TIFF, OME-TIFF, OME-Zarr, PNG, etc.) are supported out of the box via: - `bioio-ome-tiff` - OME-TIFF files - `bioio-ome-zarr` - OME-Zarr files - `bioio-tifffile` - General TIFF files - `bioio-imageio` - PNG, JPEG, and other common formats If your image format is not supported by the default readers, then you will get a warning and (by default in napari) a widget to install the suggested reader. If you know of your additional proprietary formats, install the appropriate bioio reader. See the [bioio documentation](https://bioio-devs.github.io/bioio/) for the full list of available readers. **Note**: The use of `bioio-bioformats` requires an automatic, initial download of required Java files, this takes some time. Most native format readers do a better job reading metadata compared to `bioio-bioformats`. Please [file an issue] if you encounter problems with image reading! ## Usage ### In napari Simply drag and drop image files into napari. `ndevio` handles the rest! To learn more about the decisions that ndevio (and its settings) makes when loading images, see [How ndevio Handles Images](#how-ndevio-handles-images) below. #### Multi-scene Images When opening multi-scene files (e.g., multi-position acquisitions, mosaics), a **Scene Widget** appears in the viewer, allowing you to select which scene to display. Configure default behavior via the Settings widget. #### Bioio Reader Plugin Installation Widget If you open a file that requires a bioio reader not currently installed, ndevio will display a **BioIO Plugin Installation widget** in napari suggesting the appropriate plugin to install. ![bioio plugin installation suggestion widget](https://github.com/ndev-kit/ndevio/blob/main/resources/bioio-plugin-install-widget.png?raw=true) This widget taps into the `napari-plugin-manager` to install the bioio reader plugin from PyPI via a GUI. You may invoke this widget manually at any time via `Plugins > ndevio > Install BioIO Reader Plugins` to install any additional bioio reader plugin *and* update any currently installed plugins. #### Settings Widget Access **ndevio settings** via `Plugins > ndev-settings > Settings` to customize: - **Preferred reader**: Override bioio's default plugin selection priority (useful for formats with multiple compatible readers) - **Multi-scene handling**: Choose whether to show the scene widget, view all scenes as a stack, or view only the first scene - **Plugin suggestions**: Enable/disable automatic plugin installation prompts for unsupported formats ![ndevio settings via the ndev-settings widget in napari](https://github.com/ndev-kit/ndevio/blob/main/resources/ndev-settings.png?raw=true) These settings are managed by [ndev-settings] and persist across napari sessions. #### Utilities Widget Access via `Plugins > ndevio > Utilities` for: - **Batch concatenation** of images - **Metadata management** - **Export** as OME-TIFF or figure (PNG) **Note**: Elements of this widget are being superseded by [napari-metadata] for more comprehensive metadata handling. **Note 2:** This widget was built during napari-ndev mono-repo development and does not fully reflect the design goals of ndevio. It will remain functional, but expect future versions to look different. ### Programmatic Usage with `nImage` The `nImage` class extends [bioio]'s `BioImage` with napari-specific functionality: ```python from ndevio import nImage from napari import Viewer # Load image with automatic metadata extraction img = nImage("path/to/image.czi") # Because nImage subclasses BioImage, all BioImage methods are available print(img.dims) # e.g., <Dimensions [T: 15, C: 4, Z: 1, Y: 256, X: 256]> # Access napari-ready properties, note that channel and singleton dimensions are dropped print(img.reference_xarray) # e.g., xarray.DataArray with dims (T, Y, X) and shape (15, 256, 256) print(img.layer_data[0]) # e.g. the highest resolution array; a list is returned for multiscale image support print(img.layer_scale) # e.g., (1.0, 0.2, 0.2) - time interval + physical scale per dimension, napari ready print(img.layer_axis_labels) # e.g., ('T', 'Y', 'X') print(img.layer_units) # e.g., ('s', 'µm', 'µm') print(img.layer_metadata) # e.g., a dictionary containing the 1) full BioImage object, 2) raw_image metadata and 3) OME metadata (if parsed) - accessible via `viewer.layers[n].metadata` # A convenience method to get napari LayerDataTuples with nImage metadata for napari viewer = Viewer() for ldt in img.get_layer_data_tuples(): viewer.add_layer(ldt) ``` ### Sample Data ndevio includes sample datasets accessible via `File > Open Sample > ndevio`. These samples use the `nImage` API to demonstrate the metadata handling. ![4 Channel 2D neural cells with segmentation labels](https://github.com/ndev-kit/ndevio/blob/main/resources/neuron-sampledata.png?raw=true) - 2D neural cells in a dish imaged with 4 channels, with corresponding segmentation labels - 2D brain slice with 3 different transcription factor antibody stains - A single 2D+Time sample from a scratched retinal epithelial cell culture, including auto-detected labels in the same file. - the napari-ndev logo as a .png ## How ndevio Handles Images ### Metadata Image metadata is extracted from bioio and converted to napari layer metadata based on the current napari convention to squeeze out singleton dimensions (i.e. drop dimensions of size 1): - **Full metadata** available at `viewer.layers[n].metadata` - **Time and physical scale** automatically applied to layers from OME metadata or file headers - **Dimension labels** (T, C, Z, Y, X) preserved in `axis_labels` - **Physical units** (µm, nm, etc.) stored in layer metadata ### Memory Management Images are loaded **in-memory** or **lazily** (via dask) automatically based on: - File size < 4 GB **AND** - File size < 30% of available RAM - Remote files (e.g., S3, HTTP) and multiscale are always loaded lazily ### Multi-channel Images Multi-channel images are **always split** into individual layers (one per channel), using channel names from metadata when available. Images are added with colorblind-friendly colormaps. ### Mosaic/Tiled Images Images with tiles (e.g., stitched acquisitions) are **automatically stitched together** if the reader supports this behavior. ### RGB Images RGB(A) images are currently added to the viewer as a single RGB(A) layer, according to napari conventions. RGB(A) images are identified by containing the Samples ('S') dimension by bioio, which nominally exists for images with the last (`-1`) dimension being of size 3 or 4. ### Detection of Labels/Segmentation Layers If an image contains a channel name or file name suggestive of a labels layer, ndevio will add that channel as a labels layer in napari. Mixed image and label files are possible by having information in the channel names (e.g., `["DAPI", "DAPI-labels"]`). ### Customization If you need different behavior for any of these automated handling rules, please [file an issue] — we may be able to add settings to configure them! ## Coming Soon **Writers for OME-TIFF and OME-Zarr** with round-trip napari metadata support! ## Contributing Contributions are very welcome! Please see [Contributing.md](CONTRIBUTING.md) for development setup and guidelines. ## License Distributed under the terms of the [BSD-3] license, "ndevio" is free and open source software ## Issues If you encounter any problems, please [file an issue] along with a detailed description. [file an issue]: https://github.com/ndev-kit/ndevio/issues [napari]: https://github.com/napari/napari [copier]: https://copier.readthedocs.io/en/stable/ [BSD-3]: http://opensource.org/licenses/BSD-3-Clause [napari-plugin-template]: https://github.com/napari/napari-plugin-template [PyPI]: https://pypi.org/project/ndevio/ [tox]: https://tox.readthedocs.io/en/latest/ [bioio]: https://github.com/bioio-devs/bioio [napari-aicsimageio]: https://github.com/AllenCellModeling/napari-aicsimageio [ndev-settings]: https://github.com/ndev-kit/ndev-settings [napari-metadata]: https://github.com/napari/napari-metadata [nbatch]: https://github.com/ndev-kit/nbatch [uv]: https://docs.astral.sh/uv/ [ndev-kit]: https://github.com/ndev-kit [bioio-ome-zarr]: https://github.com/bioio-devs/bioio-ome-zarr
text/markdown
Tim Monko
timmonko@gmail.com
null
null
null
null
[ "Development Status :: 2 - Pre-Alpha", "Framework :: napari", "Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Scientific/Engineering :: Image Processing" ]
[]
null
null
>=3.11
[]
[]
[]
[ "bioio-base", "bioio-imageio", "bioio-ome-tiff>=1.2.0", "bioio-ome-zarr>=3", "bioio-tifffile", "bioio>=3.2.0", "magic-class", "magicgui", "napari", "napari-plugin-manager>=0.1.7", "natsort", "nbatch>=0.0.4", "ndev-settings>=0.4.1", "pooch", "xarray", "zarr>=3.1.3", "napari[all]; extra == \"all\"" ]
[]
[]
[]
[ "Bug Tracker, https://github.com/ndev-kit/ndevio/issues", "Documentation, https://github.com/ndev-kit/ndevio#README.md", "Source Code, https://github.com/ndev-kit/ndevio", "User Support, https://github.com/ndev-kit/ndevio/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:35:28.108535
ndevio-0.9.0.tar.gz
8,394,855
65/77/40c08c3bf59fce9b84d9711eb6ba2fa630aaa52ab75fa7252550ff53a60b/ndevio-0.9.0.tar.gz
source
sdist
null
false
e3398be66e66abec763d009847f65b3e
e4f87611417cbdcf833cf3cb12a17b99fb2e6a54caf1a4c698aa1d16f6950653
657740c08c3bf59fce9b84d9711eb6ba2fa630aaa52ab75fa7252550ff53a60b
BSD-3-Clause
[ "LICENSE" ]
204
2.4
nano-eval
0.2.7
Nano Eval - A minimal tool for verifying VLMs/LLMs across frameworks
**nano-eval** is a minimal tool for measuring the quality of a text or vision model. ## Quickstart ```bash uvx nano-eval -m text -m vision --max-samples 100 # prints: Task Accuracy Samples Duration Output Tokens Per Req Tok/s ------ -------- ------- -------- ------------- ------------- text 86.0% 100 15s 11873 7658 vision 72.0% 100 37s 8714 1894 ``` > **Note:** This tool is for eyeballing the accuracy of a model. One use case is comparing accuracy between inference frameworks (e.g., vLLM vs SGLang vs MAX running the same model). ## Supported Modalities | Modality | Dataset | Description | |----------|---------|-------------| | `text` | gsm8k_cot_llama | Grade school math with chain-of-thought (8-shot) | | `vision` | HuggingFaceM4/ChartQA | Chart question answering with images | ## Usage ``` $ nano-eval --help Usage: nano-eval [OPTIONS] Evaluate LLMs on standardized tasks via OpenAI-compatible APIs. Example: nano-eval -m text Options: -m, --modality [text|vision] Modality to evaluate (can be repeated) [required] --base-url TEXT OpenAI-compatible API endpoint; tries 127.0.0.1:8000/8080 if omitted --model TEXT Model name; auto-detected if endpoint serves one model --api-key TEXT Bearer token for API authentication --max-concurrent INTEGER [default: 8] --extra-request-params TEXT API params as key=value,... [default: temperature=0,max_tokens=256,seed=42] --max-samples INTEGER If provided, limit samples per task --output-path PATH Write eval_results.json and request logs to this directory --log-requests Save per-request results as JSONL (requires --output-path) --dataset-seed INTEGER Controls sample order [default: 42] --request-timeout INTEGER Timeout in seconds for each API request [default: 300] -v, --verbose Increase verbosity (up to -vvv) --version Show the version and exit. --help Show this message and exit. ``` ### Python API ```python from nano_eval import evaluate result = evaluate( modalities=["text"], base_url="http://127.0.0.1:8000/v1", model="meta-llama/Llama-3.2-1B-Instruct", max_samples=100, ) print(f"Accuracy: {result['results']['text']['metrics']['accuracy']:.1%}") ``` ## Example Output When using `--output-path`, an `eval_results.json` file is generated: ```json { "config": { "max_samples": 100, "model": "deepseek-chat" }, "framework_version": "0.2.6", "results": { "text": { "elapsed_seconds": 15.51, "metrics": { "accuracy": 0.86, "accuracy_stderr": 0.03487350880197947 }, "num_samples": 100, "samples_hash": "12a1e9404db6afe810290a474d69cfebdaffefd0b56e48ac80e1fec0f286d659", "task": "gsm8k_cot_llama", "modality": "text", "total_input_tokens": 106965, "total_output_tokens": 11873, "tokens_per_second": 7658.994842036105 } }, "total_seconds": 15.51 } ``` With `--log-requests`, a `request_log_{modality}.jsonl` is written per modality: ```json { "request_id": 0, "target": "4", "prompt": "What is 2+2?", "response": "4", "score": 1.0, "stop_reason": "stop", "input_tokens": 7, "output_tokens": 1, "duration_seconds": 0.83 } ``` --- Inspired by [lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-harness).
text/markdown
null
Thomas Børstad <tboerstad@users.noreply.github.com>
null
null
null
llm, evaluation, benchmark, openai, api, vllm, tgi
[ "Development Status :: 4 - Beta", "Programming Language :: Python :: 3", "Operating System :: OS Independent", "Topic :: Scientific/Engineering :: Artificial Intelligence" ]
[]
null
null
>=3.10
[]
[]
[]
[ "click>=8.0.0", "datasets>=2.0.0", "httpx>=0.23.0", "pillow>=9.0.0", "tqdm>=4.62.0", "multiprocess<=0.70.17", "pytest>=7.0.0; extra == \"dev\"", "pytest-cov>=4.0.0; extra == \"dev\"", "ruff>=0.1.0; extra == \"dev\"", "prek>=0.2.0; extra == \"dev\"", "respx>=0.20.0; extra == \"dev\"", "ty>=0.0.1; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://github.com/tboerstad/nano-eval", "Repository, https://github.com/tboerstad/nano-eval", "Issues, https://github.com/tboerstad/nano-eval/issues" ]
uv/0.5.14
2026-02-20T21:35:24.239799
nano_eval-0.2.7.tar.gz
15,224
e1/fb/3d14357a510212a11e712682ab747ff3ad33bfaf883b79dede840c5b7d9e/nano_eval-0.2.7.tar.gz
source
sdist
null
false
58e3925aebf7b9ae2af2c0a3f7743ea7
2a84770839cd4cfba1cb777220e852b0d458cde93035dfece885e34717f9f359
e1fb3d14357a510212a11e712682ab747ff3ad33bfaf883b79dede840c5b7d9e
MIT
[ "LICENSE" ]
200
2.4
eip-search
0.3.2
CLI intelligence tool for the Exploit Intelligence Platform — a modern searchsploit replacement
# Exploit Intel Platform CLI Search Tool Package/command: `eip-search` <p align="center"> <img src="https://exploit-intel.com/static/brand/mark-cyan.svg" width="160" alt="Exploit Intel Platform (EIP)" /> </p> A modern **searchsploit replacement** powered by the [Exploit Intelligence Platform](https://exploit-intel.com). Search 370K+ vulnerabilities and 105K+ exploits from 4 sources with risk intelligence, exploit quality ranking, Nuclei scanner integration, and trojan warnings — all from your terminal. ![eip-search CLI screenshot](https://codeberg.org/exploit-intel/eip-search/raw/branch/main/eip-search.png) Part of the same project family: - [`eip-search`](https://codeberg.org/exploit-intel/eip-search) — terminal client - [`eip-mcp`](https://codeberg.org/exploit-intel/eip-mcp) — MCP server for AI assistants ## Highlights - Search 370K+ vulnerabilities and 105K+ exploits from one CLI - Browse exploits directly by source, language, vendor, or attack type - **Generate PoC exploits** for any CVE using a local LLM (Ollama) with optional vision pipeline for writeup screenshots - Download exploit code by CVE ID — interactive picker selects the best match - Combine CVSS, EPSS, KEV, and exploit quality in one view - Surface trusted exploit sources first and flag trojans clearly - Pull Nuclei templates plus Shodan/FOFA/Google recon dorks - Browse authors, CWEs, vendors, and products — resolve EDB/GHSA IDs to CVEs ## Why eip-search? **searchsploit** is grep over a CSV. It can tell you an exploit exists, but nothing about how dangerous the vulnerability is, how reliable the exploit is, or whether it's secretly a trojan. **eip-search** combines data from NVD, CISA KEV, EPSS, ExploitDB, Metasploit, GitHub, and nomi-sec into a single tool that answers questions searchsploit never could: - "What critical Fortinet vulns are being actively exploited right now?" - "Which of these 127 BlueKeep exploits is actually reliable — and which one is a trojan?" - "Give me the Shodan dork to find exposed TeamCity instances for CVE-2024-27198" ## Installation ### Requirements - **Python 3.10 or newer** (check with `python3 --version` or `python --version`) - **pip** (comes with Python on most systems) ### macOS ```bash # Install Python 3 via Homebrew (if not already installed) brew install python3 # Option 1: Virtual environment (recommended) python3 -m venv ~/.venvs/eip source ~/.venvs/eip/bin/activate pip install eip-search # Option 2: pipx (isolated, no venv activation needed) brew install pipx pipx install eip-search # The 'eip-search' command is now available eip-search --version ``` ### Kali Linux / Debian / Ubuntu ```bash # Python 3 is pre-installed on Kali. Install pip if needed: sudo apt update && sudo apt install -y python3-pip python3-venv # Option 1: Install into a virtual environment (recommended) python3 -m venv ~/.venvs/eip source ~/.venvs/eip/bin/activate pip install eip-search # Option 2: Install with pipx (isolated, no venv management) sudo apt install -y pipx pipx install eip-search # The 'eip-search' command is now available eip-search --version ``` > **Kali users**: If you see `error: externally-managed-environment`, use one of the virtual environment methods above. Kali 2024+ enforces PEP 668 which blocks global pip installs. ### Windows ```powershell # Install Python 3 from https://python.org (check "Add to PATH" during install) # Option 1: Virtual environment python -m venv %USERPROFILE%\.venvs\eip %USERPROFILE%\.venvs\eip\Scripts\activate pip install eip-search # Option 2: pipx pip install pipx pipx install eip-search # The 'eip-search' command is now available eip-search --version ``` > **Windows Terminal** or **PowerShell** is recommended for full color and Unicode support. The classic `cmd.exe` may not render tables correctly. ### Arch Linux / Manjaro ```bash sudo pacman -S python python-pip python-pipx pipx install eip-search ``` ### From Source (all platforms) ```bash git clone git@codeberg.org:exploit-intel/eip-search.git cd eip-search python3 -m venv .venv source .venv/bin/activate # Linux/macOS # .venv\Scripts\activate # Windows pip install -e . ``` ## Building Packages ### Build Dependencies | Target | Requirements | |---|---| | `make build` | Python 3, `build` module (`pip install build`) | | `make check` / `make pypi` | `twine` (`pip install twine`) | | `make deb` | Docker | | `make tag-release` | Python 3 (version bump only — Forgejo Actions handles the rest) | | `make release` | All of the above + `tea` CLI ([codeberg.org/gitea/tea](https://codeberg.org/gitea/tea)) | Install everything at once: ```bash pip install build twine # Docker: https://docs.docker.com/get-docker/ # tea CLI: https://codeberg.org/gitea/tea ``` The Makefile checks for each dependency before running and will tell you exactly what's missing. ### PyPI (wheel + sdist) ```bash make build # build dist/*.whl and dist/*.tar.gz make check # validate with twine make pypi # upload to PyPI ``` ### .deb Packages Build for a single distro or all four supported targets: ```bash make deb DISTRO=ubuntu-jammy # Ubuntu 22.04 make deb DISTRO=ubuntu-noble # Ubuntu 24.04 make deb DISTRO=debian-bookworm # Debian 12 make deb DISTRO=kali # Kali Rolling make deb # all four ``` Output lands in `dist/`: ``` dist/eip-search_0.2.0_ubuntu-jammy_all.deb dist/eip-search_0.2.0_ubuntu-noble_all.deb dist/eip-search_0.2.0_debian-bookworm_all.deb dist/eip-search_0.2.0_kali-rolling_all.deb ``` ### Releasing **One-time setup:** add `PYPI_API_TOKEN` and `RELEASE_TOKEN` as repository secrets in Codeberg (Settings → Actions → Secrets). **Automated release (recommended)** — bumps version, commits, tags, and pushes. Forgejo Actions builds PyPI packages + all 4 `.deb`s, uploads to PyPI, and creates a Codeberg release with artifacts attached: ```bash make tag-release VERSION=0.2.0 ``` **Local release (alternative)** — does everything locally without CI: ```bash make release VERSION=0.2.0 ``` ### Shell Completion (optional) Enable tab completion for your shell (run from an interactive terminal): ```bash # Bash eip-search --install-completion bash # Zsh eip-search --install-completion zsh # Fish eip-search --install-completion fish # PowerShell eip-search --install-completion powershell ``` ### Verify Installation ```bash eip-search --version # eip-search 0.1.4 eip-search stats # Should display platform statistics if your network can reach exploit-intel.com ``` ### Troubleshooting | Problem | Solution | |---|---| | `command not found: eip-search` | Make sure your virtual environment is activated, or use `pipx` which manages PATH automatically | | `externally-managed-environment` | Use a virtual environment or `pipx` — see instructions above | | `SSL certificate error` | Your Python may lack certificates. On macOS: `brew reinstall python3`. On Linux: `sudo apt install ca-certificates` | | `Connection refused` / timeouts | Check that you can reach `https://exploit-intel.com` — the tool requires internet access | | Tables look broken | Use a terminal with Unicode support (Windows Terminal, iTerm2, any modern Linux terminal) | ## Quick Start The simplest usage mirrors searchsploit — just type what you're looking for: ``` $ eip-search "palo alto" ``` ``` ┏━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━┳━━━━━━━┳━━━━━━━━┳━━━━━━┳━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┓ ┃CVE ┃ Sev ┃ CVSS ┃ EPSS ┃ Exp ┃ ┃ Title ┃ ┡━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━╇━━━━━━━╇━━━━━━━━╇━━━━━━╇━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┩ │CVE-2025-0108 │ CRITICAL │ 9.1 │ 94.0% │ 16 │ KEV │ Palo Alto Networks PAN-OS … │ │CVE-2025-0107 │ CRITICAL │ 9.8 │ 77.0% │ 1 │ │ Palo Alto Networks Expedi… │ │CVE-2025-0111 │ MEDIUM │ 6.5 │ 2.0% │ 2 │ KEV │ Palo Alto Networks PAN-OS … │ │ ... │ │ │ │ │ │ │ └─────────────────┴────────────┴───────┴────────┴──────┴─────┴──────────────────────────────┘ Page 1/9 (41 total results) ``` Every result includes CVSS score, EPSS exploitation probability, exploit count, and CISA KEV status — context searchsploit simply doesn't have. ## CVE Intelligence Briefs Type a CVE ID and get a full intelligence brief — no subcommand needed: ``` $ eip-search CVE-2024-3400 ``` ``` ╭──────────────────────────────╮ │ CVE-2024-3400 CRITICAL KEV │ ╰──────────────────────────────╯ Palo Alto Networks PAN-OS Unauthenticated Remote Code Execution CVSS: 10.0 (CVSS:3.1/AV:N/AC:L/PR:N/UI:N/S:C/C:H/I:H/A:H) EPSS: 94.3% (99.9th percentile) Attack Vector: NETWORK | CWE: CWE-77, CWE-20 | Published: 2024-04-12 | KEV added: 2024-04-12 A command injection as a result of arbitrary file creation vulnerability in the GlobalProtect feature of Palo Alto Networks PAN-OS software ... Affected Products - paloaltonetworks/pan-os ... and 40 more Exploits (43) MODULES #48006 metasploit ruby panos_telemetry_cmd_exec.rb Rank: excellent LLM: working_poc has code PROOF OF CONCEPT #9546 exploitdb text EDB-51996 LLM: working_poc has code #370108 ★ 161 github http h4x0r-dz/CVE-2024-3400 LLM: working_poc has code #369757 ★ 90 github python W01fh4cker/CVE-2024-3400-RCE-Scan LLM: working_poc has code #369206 ★ 72 github python 0x0d3ad/CVE-2024-3400 LLM: working_poc has code ... ... and 32 more PoCs (use --all to show) Tip: eip-search view <id> | eip-search download <id> -x Also Known As - EDB: EDB-51996 - GHSA: GHSA-v475-xhc9-wfxg References - [Vendor Advisory] https://security.paloaltonetworks.com/CVE-2024-3400 - [Exploit, Vendor Advisory] https://unit42.paloaltonetworks.com/cve-2024-3400/ ... ``` Exploits are **grouped by quality** (Metasploit modules first, then verified ExploitDB, then GitHub PoCs ranked by stars) and **ranked by a composite score**. ## Trojan Detection BlueKeep (CVE-2019-0708) has 127 exploits. One of them is a trojan. eip-search warns you: ``` $ eip-search info CVE-2019-0708 ``` ``` ╭──────────────────────────────╮ │ CVE-2019-0708 CRITICAL KEV │ ╰──────────────────────────────╯ CVE-2019-0708 BlueKeep RDP Remote Windows Kernel Use After Free CVSS: 9.8 EPSS: 94.5% (100.0th percentile) Exploits (127) MODULES #47841 metasploit ruby cve_2019_0708_bluekeep_rce.rb Rank: manual LLM: working_poc has code #47840 metasploit ruby cve_2019_0708_bluekeep.rb LLM: working_poc has code VERIFIED #9123 exploitdb ruby EDB-47416 LLM: working_poc ✓ verified has code PROOF OF CONCEPT #72412 ★ 1187 nomisec Ekultek/BlueKeep #72419 ★ 497 nomisec n1xbyte/CVE-2019-0708 #72417 ★ 389 nomisec k8gege/CVE-2019-0708 ... ... and 113 more PoCs (use --all to show) SUSPICIOUS #72431 ★ 2 nomisec ttsite/CVE-2019-0708- ⚠ TROJAN — flagged by AI analysis Tip: eip-search view <id> | eip-search download <id> -x ``` The Metasploit modules and verified ExploitDB entry surface to the top. The trojan sinks to the bottom with a clear warning. ## Risk-Based Triage "What critical Fortinet vulnerabilities with public exploits should I worry about right now?" ``` $ eip-search triage --vendor fortinet --severity critical ``` ``` TRIAGE — vulnerabilities with exploits, sorted by exploitation risk Filters: vendor=fortinet, severity=critical, EPSS>=0.5 ┏━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━┳━━━━━━━┳━━━━━━━━┳━━━━━━┳━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┓ ┃CVE ┃ Sev ┃ CVSS ┃ EPSS ┃ Exp ┃ ┃ Title ┃ ┡━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━╇━━━━━━━╇━━━━━━━━╇━━━━━━╇━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┩ │CVE-2018-13379 │ CRITICAL │ 9.1 │ 94.5% │ 14 │ KEV │ Fortinet FortiProxy Path … │ │CVE-2022-40684 │ CRITICAL │ 9.8 │ 94.4% │ 30 │ KEV │ Fortinet FortiProxy Auth … │ │CVE-2023-48788 │ CRITICAL │ 9.8 │ 94.2% │ 1 │ KEV │ Fortinet FortiClient SQL … │ │CVE-2024-55591 │ CRITICAL │ 9.8 │ 94.2% │ 8 │ KEV │ Fortinet FortiProxy Auth … │ │CVE-2022-42475 │ CRITICAL │ 9.8 │ 94.0% │ 7 │ KEV │ Fortinet FortiOS Buffer … │ └─────────────────┴────────────┴───────┴────────┴──────┴─────┴──────────────────────────────┘ Page 1/1 (17 total results) ``` Triage defaults to showing vulnerabilities with public exploits and EPSS >= 0.5, sorted by exploitation probability. Every result here is confirmed actively exploited (KEV), has dozens of public exploits, and has a >94% chance of being exploited in the wild. ## Nuclei Templates & Recon Dorks Get scanner templates with ready-to-paste Shodan, FOFA, and Google dorks: ``` $ eip-search nuclei CVE-2024-27198 ``` ``` ╭──────────────────────────────────╮ │ CVE-2024-27198 Nuclei Templates │ ╰──────────────────────────────────╯ TeamCity < 2023.11.4 - Authentication Bypass Nuclei Templates (1) CVE-2024-27198 ✓ verified critical TeamCity < 2023.11.4 - Authentication Bypass Author: DhiyaneshDk Tags: cve, cve2024, teamcity, jetbrains, auth-bypass, kev, vkev, vuln Recon Queries: Shodan: http.component:"TeamCity" || http.title:teamcity || http.component:"teamcity" FOFA: title=teamcity Google: intitle:teamcity Run: nuclei -t CVE-2024-27198 -u https://target.com ``` ## Browse Exploits Search exploits directly by source, language, vendor, author, or attack type — no CVE ID needed: ```bash # All Metasploit RCE modules eip-search exploits --source metasploit --attack-type RCE # Python exploits for Fortinet with downloadable code eip-search exploits "fortinet" --language python --has-code # Exploits for a specific CVE eip-search exploits --cve CVE-2024-3400 # Exploits by a specific author, ranked by GitHub stars eip-search exploits --author "Chocapikk" --sort stars_desc ``` ``` $ eip-search exploits "mitel" --has-code -n 5 ``` ``` ┏━━━━━━━━━┳━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━┳━━━━━━━━━━━━━┳━━━━━━━━┳━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━━┓ ┃ ID ┃ CVE ┃ Sev ┃ Source ┃ Lang ┃ ★ ┃ Name ┃ ┡━━━━━━━━━╇━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━╇━━━━━━━━━━━━━╇━━━━━━━━╇━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━━┩ │ 426906 │ CVE-2024-41713 │ CRITICAL │ nomisec │ │ 19 │ watchtowrlabs/Mitel-M… │ │ 426908 │ CVE-2024-41713 │ CRITICAL │ nomisec │ │ │ Sanandd/cve-2024-CVE… │ │ 426907 │ CVE-2024-41713 │ CRITICAL │ nomisec │ │ │ zxj-hub/CVE-2024-417… │ │ 426615 │ CVE-2024-35315 │ MEDIUM │ nomisec │ │ 1 │ ewilded/CVE-2024-353… │ │ 426909 │ CVE-2024-41713 │ CRITICAL │ nomisec │ │ │ amanverma-wsu/CVE-20… │ └─────────┴──────────────────┴────────────┴─────────────┴────────┴─────┴─────────────────────────┘ Page 1/5 (89 total results) Tip: eip-search view <id> | eip-search download <id> -x ``` Every result includes the exploit ID, associated CVE, severity, source, language, and GitHub stars. Use the exploit ID directly with `view` or `download`. ## Reference Data Browse authors, CWEs, vendors, and products, or resolve alternate identifiers to CVEs: ```bash # Top exploit authors eip-search authors # Author profile with their exploits eip-search author Metasploit eip-search author "Chocapikk" --page 2 # CWE categories ranked by vuln count eip-search cwes # CWE detail eip-search cwe 79 eip-search cwe CWE-89 # Top vendors by vulnerability count eip-search vendors # Products for a vendor (discover exact CPE names for filtering) eip-search products apache eip-search products microsoft # Resolve ExploitDB or GHSA ID to its CVE eip-search lookup EDB-45961 eip-search lookup GHSA-jfh8-c2jp-5v3q ``` The `products` command is especially useful for discovering exact product names to use with `--product` filters. Product names follow CPE conventions (e.g. `http_server` not `apache httpd`, `exchange_server` not `exchange`). ## View Exploit Source Code Read exploit code directly in your terminal with syntax highlighting. Pass an exploit ID or a CVE ID: ```bash # By exploit ID (from search, info, or exploits output) $ eip-search view 77423 # By CVE ID — shows an interactive picker to choose which exploit $ eip-search view CVE-2024-3400 ``` ``` Exploits for CVE-2024-3400: [1] #48006 metasploit ruby panos_telemetry_cmd_exec.rb Rank: excellent working_poc [2] #9546 exploitdb text EDB-51996 working_poc [3] #370108 ★ 161 github http h4x0r-dz/CVE-2024-3400 working_poc Select [1-43, default=1]: 1 ``` ``` panos_telemetry_cmd_exec.rb 1 ## 2 # This module requires Metasploit: https://metasploit.com/download 3 # Current source: https://github.com/rapid7/metasploit-framework 4 ## 5 6 class MetasploitModule < Msf::Exploit::Remote 7 Rank = ExcellentRanking 8 ... ``` When an exploit has multiple files, eip-search auto-selects the most relevant code file. Use `--file` to pick a specific one. ## Download Exploit Code Download and optionally extract exploit archives. Pass an exploit ID or a CVE ID: ```bash # By CVE ID — interactive picker, auto-extracts $ eip-search download CVE-2024-3400 --extract ``` ``` Exploits with code for CVE-2024-3400: [1] #48006 metasploit ruby panos_telemetry_cmd_exec.rb Rank: excellent working_poc [2] #9546 exploitdb text EDB-51996 working_poc ... Select [1-43, default=1]: 1 Downloaded: metasploit-modules_exploits_linux_http_panos_telemetry_cmd_exec.rb.zip ZIP password: eip (exploit archives are password-protected to prevent AV quarantine) Extracted: metasploit-modules_exploits_linux_http_panos_telemetry_cmd_exec.rb/ Files (1): - panos_telemetry_cmd_exec.rb ``` ```bash # By exploit ID — downloads directly, no picker $ eip-search download 77423 --extract ``` ``` Downloaded: nomisec-fullhunt_log4j-scan.zip ZIP password: eip (exploit archives are password-protected to prevent AV quarantine) Extracted: nomisec-fullhunt_log4j-scan/ Files (10): - fullhunt-log4j-scan-07f7e32/.gitignore - fullhunt-log4j-scan-07f7e32/Dockerfile - fullhunt-log4j-scan-07f7e32/log4j-scan.py - fullhunt-log4j-scan-07f7e32/requirements.txt ... ``` > **Note:** Downloaded ZIPs are encrypted with password **`eip`** as a safety measure to prevent antivirus software from quarantining exploit code. Use `--extract` / `-x` to automatically unzip. ## Advanced Search The `search` subcommand exposes the full filter set: ```bash # All SQL injection vulns with public exploits, sorted by CVSS eip-search search --cwe 89 --has-exploits --sort cvss_desc # Critical KEV entries with high exploitation probability eip-search search --kev --severity critical --min-epss 0.9 # Recent npm vulnerabilities with exploits eip-search search --ecosystem npm --has-exploits --sort newest # Microsoft Exchange critical vulns eip-search search --product exchange --severity critical --has-exploits ``` ``` $ eip-search search --cwe 89 --has-exploits --sort cvss_desc -n 5 ``` ``` ┏━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━┳━━━━━━━┳━━━━━━━━┳━━━━━━┳━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┓ ┃CVE ┃ Sev ┃ CVSS ┃ EPSS ┃ Exp ┃ ┃ Title ┃ ┡━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━╇━━━━━━━╇━━━━━━━━╇━━━━━━╇━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┩ │CVE-2024-3605 │ CRITICAL │ 10.0 │ 64.9% │ 1 │ │ Thimpress WP Hotel Booking… │ │CVE-2024-3922 │ CRITICAL │ 10.0 │ 88.5% │ 3 │ │ Dokan Pro Plugin SQL Inje… │ │CVE-2024-39911 │ CRITICAL │ 10.0 │ 68.3% │ 1 │ │ Fit2cloud 1panel SQL Inje… │ │CVE-2025-52694 │ CRITICAL │ 10.0 │ 9.7% │ 1 │ │ Advantech IoT Edge SQL In… │ │CVE-2024-43918 │ CRITICAL │ 10.0 │ 48.9% │ 1 │ │ Woobewoo Product Table SQ… │ └─────────────────┴────────────┴───────┴────────┴──────┴─────┴──────────────────────────────┘ Page 1/817 (4,082 total results) ``` ## JSON Output for Scripting Every command supports `--json` for piping into `jq`, scripts, or SIEMs: ``` $ eip-search search "log4j" --has-exploits --sort epss_desc -n 5 --json ``` ```json { "total": 15, "page": 1, "per_page": 5, "total_pages": 3, "items": [ { "cve_id": "CVE-2021-44228", "title": "Log4Shell HTTP Header Injection", "severity_label": "critical", "cvss_v3_score": 10.0, "epss_score": 0.94358, "is_kev": true, "exploit_count": 401 }, ... ] } ``` ```bash # Get all critical KEV CVE IDs as a flat list eip-search search --kev --severity critical -n 100 --json | jq -r '.items[].cve_id' # Feed into nuclei eip-search search --has-nuclei --severity critical --json | jq -r '.items[].cve_id' | xargs -I{} nuclei -t {} -u https://target.com ``` ## Platform Statistics ``` $ eip-search stats ``` ``` ╭───────────────────────────────╮ │ Exploit Intelligence Platform │ ╰───────────────────────────────╯ ┌──────────────────────────────┬─────────────────────┐ │ Total Vulnerabilities │ 370,791 │ │ Published │ 191,380 │ │ With CVSS Scores │ 238,607 │ │ With EPSS Scores │ 315,656 │ │ Critical Severity │ 29,145 │ │ CISA KEV Entries │ 1,522 │ │ │ │ │ Vulns with Exploits │ 90,481 │ │ Total Exploits │ 105,731 │ │ With Nuclei Templates │ 404 │ │ │ │ │ Vendors Tracked │ 37,508 │ │ Exploit Authors │ 23,281 │ │ │ │ │ Last Updated │ 2026-02-17 23:07:26 │ └──────────────────────────────┴─────────────────────┘ ``` ## Generate Exploits with Local LLM Generate a proof-of-concept exploit for any CVE using a local [Ollama](https://ollama.com) instance. The tool fetches all available intelligence from the platform — writeup text, existing exploit code, and screenshots — then uses a two-stage LLM pipeline to produce a clean, minimal Python PoC. ```bash # Check feasibility first (no Ollama needed) $ eip-search generate CVE-2026-2686 --check ``` ``` CVE-2026-2686 — SECCN Dingcheng G10 Command Injection CVSS 9.8 | RCE | trivial | Feasibility: EXCELLENT (11) Reasons: web-based (RCE), trivial, has writeup, HTTP details in summary, known CWE pattern Files: 1 text, 8 screenshots ``` ```bash # Generate the exploit $ eip-search generate CVE-2026-2686 -o exploit.py ``` ``` CVE-2026-2686 — SECCN Dingcheng G10 Command Injection CVSS 9.8 | RCE | trivial | Feasibility: EXCELLENT (11) Analyzing 8 screenshots... img-001: telnet session, root shell on BusyBox (8.9s) img-003: POST /cgi-bin/session_login.cgi with injection payload (10.7s) img-008: Burp capture with full request headers (16.9s) 2 screenshots skipped (no actionable details) Generating PoC with kimi-k2:1t-cloud... done (11s) (syntax-highlighted Python exploit) Saved: exploit.py ``` The generator works in three modes depending on what's available: - **Writeup + screenshots** → vision model extracts technical details from images, code model generates PoC from enriched context - **Existing exploit code** → code model rewrites/fixes it into a clean, standardized Python PoC - **CVE description only** → generates from NVD description and LLM analysis (lower quality) Generated exploits are minimal proofs of concept (inject `id` for RCE, extract `@@version` for SQLi) — no backdoors, reverse shells, or weaponization. Each script is clearly marked as LLM-generated and untested. **Requirements:** [Ollama](https://ollama.com) running locally with a code model pulled. Vision model is optional (used for screenshot analysis). ```bash # Install Ollama, then pull models ollama pull kimi-k2:1t-cloud # code generation (required) ollama pull qwen3-vl:235b-instruct-cloud # screenshot analysis (optional) ``` ```bash # Options eip-search generate CVE-ID # full pipeline (vision + code) eip-search generate CVE-ID --check # feasibility check only eip-search generate CVE-ID --no-vision # skip screenshots (faster) eip-search generate CVE-ID -m glm-5:cloud # override code model eip-search generate CVE-ID -o exploit.py # save to file ``` Configure defaults in `~/.eip-search.toml`: ```toml [generate] ollama_url = "http://127.0.0.1:11434" code_model = "kimi-k2:1t-cloud" vision_model = "qwen3-vl:235b-instruct-cloud" ``` ## All Commands | Command | Description | |---|---| | `eip-search "query"` | Quick search (auto-routes CVE IDs to detail view) | | `eip-search search "query" [filters]` | Search vulnerabilities with full filter support | | `eip-search exploits "query" [filters]` | Browse/search exploits directly | | `eip-search info CVE-ID` | Full intelligence brief for a vulnerability | | `eip-search generate CVE-ID` | Generate a PoC exploit using local LLM (requires Ollama) | | `eip-search triage [filters]` | Risk-sorted view of what to worry about | | `eip-search nuclei CVE-ID` | Nuclei templates + Shodan/FOFA/Google dorks | | `eip-search view ID-or-CVE` | Syntax-highlighted exploit source code | | `eip-search download ID-or-CVE` | Download exploit code as ZIP | | `eip-search stats` | Platform-wide statistics | | `eip-search authors` | Top exploit authors ranked by exploit count | | `eip-search author NAME` | Author profile with their exploits | | `eip-search cwes` | CWE categories ranked by vulnerability count | | `eip-search cwe ID` | CWE detail (accepts `79` or `CWE-79`) | | `eip-search vendors` | Top vendors ranked by vulnerability count | | `eip-search products VENDOR` | Products for a vendor (discover CPE names for filtering) | | `eip-search lookup ALT-ID` | Resolve EDB/GHSA identifier to CVE | The `view` and `download` commands accept either an exploit ID (e.g. `77423`) or a CVE ID (e.g. `CVE-2024-3400`). When given a CVE, they show an interactive picker ranked by exploit quality. ## Search Filters | Filter | Short | Description | |---|---|---| | `--severity` | `-s` | critical, high, medium, low | | `--has-exploits` | `-e` | Only CVEs with public exploit code | | `--kev` | `-k` | Only CISA Known Exploited Vulnerabilities | | `--has-nuclei` | | Only CVEs with Nuclei scanner templates | | `--vendor` | `-v` | Filter by vendor name | | `--product` | `-p` | Filter by product name | | `--ecosystem` | | npm, pip, maven, go, crates | | `--cwe` | | CWE ID (e.g. `79` or `CWE-79`) | | `--year` | `-y` | CVE publication year | | `--min-cvss` | | Minimum CVSS score (0-10) | | `--min-epss` | | Minimum EPSS score (0-1) | | `--date-from` | | Start date (YYYY-MM-DD) | | `--date-to` | | End date (YYYY-MM-DD) | | `--sort` | | newest, oldest, cvss_desc, epss_desc, relevance | | `--json` | `-j` | JSON output for scripting | ## Exploit Filters The `exploits` command has its own filter set for exploit-centric searching: | Filter | Short | Description | |---|---|---| | `--source` | | github, metasploit, exploitdb, nomisec | | `--language` | `-l` | python, ruby, go, c, etc. | | `--classification` | | LLM class: working_poc, scanner, trojan | | `--attack-type` | | RCE, SQLi, XSS, DoS, LPE, auth_bypass, info_leak | | `--complexity` | | trivial, simple, moderate, complex | | `--reliability` | | reliable, unreliable, untested | | `--author` | | Filter by exploit author name | | `--min-stars` | | Minimum GitHub stars | | `--has-code` | `-c` | Only exploits with downloadable code | | `--cve` | | Filter by CVE ID | | `--vendor` | `-v` | Filter by vendor name | | `--product` | `-p` | Filter by product name | | `--sort` | | newest, stars_desc | | `--json` | `-j` | JSON output for scripting | The positional query is auto-detected: CVE IDs map to `--cve`, other text maps to `--vendor`. ## How Exploit Ranking Works When a CVE has dozens or hundreds of exploits, eip-search ranks them by quality so the best ones surface first: | Source | Base Score | Why | |---|---|---| | Metasploit (`excellent`) | 1000 | Peer-reviewed, maintained by Rapid7 | | Metasploit (other ranks) | 500-900 | Still curated and tested | | ExploitDB (verified) | 550 | Human-verified by Offsec | | ExploitDB (unverified) | 300 | Published but not verified | | nomi-sec / GitHub | log10(stars) * 100 + bonus | Community signal via GitHub stars | On top of the base score, LLM classification modifiers apply: `working_poc` gets +100, `scanner` gets +50, while `trojan` gets -9999 (always last, with a warning). Exploit sources are ExploitDB (~88K), nomi-sec (~11K), Metasploit (~3.3K), and GitHub (~2.2K). ## Configuration Optional config at `~/.eip-search.toml`: ```toml [api] base_url = "https://exploit-intel.com" api_key = "your-key-here" # optional, for higher rate limits [display] per_page = 20 # default results per page ``` No API key is required. The public API allows 60 requests/minute. ## Security - **ZIP Slip protection**: All ZIP extraction paths are validated against directory traversal attacks - **Filename sanitization**: Download filenames are stripped of path components and special characters - **Download size cap**: 50 MB hard limit prevents memory exhaustion from malicious responses - **Markup injection prevention**: All API data is escaped before terminal rendering - **TLS verification**: All connections use standard certificate verification ## License MIT
text/markdown
Exploit Intelligence Platform
null
null
null
null
security, exploits, vulnerability, cve, searchsploit, pentesting
[ "Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Information Technology", "Topic :: Security", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12" ]
[]
null
null
>=3.10
[]
[]
[]
[ "typer>=0.9.0", "shellingham>=1.5.0", "rich>=13.0.0", "httpx>=0.27.0", "tomli>=2.0.1; python_version < \"3.11\"" ]
[]
[]
[]
[ "Homepage, https://exploit-intel.com", "Repository, https://codeberg.org/exploit-intel/eip-search" ]
twine/6.2.0 CPython/3.12.12
2026-02-20T21:35:15.341375
eip_search-0.3.2.tar.gz
61,441
3e/09/ad9447793a29ce5bad2e96f0390fb150b8b5b58bb116919c869a11aa041c/eip_search-0.3.2.tar.gz
source
sdist
null
false
3cb2106b5034b44a7d050b0272748477
968b1233229370620e4f66d340ec579f76b4b9580d09f8c4b56def55830ab22d
3e09ad9447793a29ce5bad2e96f0390fb150b8b5b58bb116919c869a11aa041c
MIT
[ "LICENSE" ]
220
2.4
apantias
2.0.4
Tools for analysis of data from the DANAE experiment on HEPHY
See GitHub wiki.
text/markdown
null
Florian Heinrich <florianheinrich@gmx.at>
null
null
null
null
[ "Programming Language :: Python :: 3", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent" ]
[]
null
null
>=3.13
[]
[]
[]
[ "h5py>=3.15.1", "matplotlib>=3.10.8", "numba>=0.64.0", "numpy>=2.4.2", "scikit-learn>=1.8.0", "scipy>=1.17.0", "seaborn>=0.13.2", "tables>=3.10.2" ]
[]
[]
[]
[ "Homepage, https://github.com/shakamaran/apantias", "Issues, https://github.com/shakamaran/apantias/issues" ]
uv/0.10.4 {"installer":{"name":"uv","version":"0.10.4","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
2026-02-20T21:34:55.409823
apantias-2.0.4-py3-none-any.whl
30,688
8f/21/d0cb9fe080e1f469ae6a11c22e07655d1d7cb44ad9f8d2237f1c56189b4b/apantias-2.0.4-py3-none-any.whl
py3
bdist_wheel
null
false
b16423265ae3c9eebcc087949e4ac154
699611f0ebd6c001ab97d377484e86c75922a14d0ef8979642abb3d044093e23
8f21d0cb9fe080e1f469ae6a11c22e07655d1d7cb44ad9f8d2237f1c56189b4b
null
[ "LICENSE" ]
196
2.4
oss-maintainer-toolkit
0.4.0
OSS Maintainer Toolkit — automated triage for PRs, issues, contributors, and review queues
# OSS Maintainer Toolkit Automated triage for PRs, issues, contributors, and review queues. A free GitHub Action and CLI built on a three-tier pipeline: embedding-based dedup, heuristic scoring, and optional LLM vision alignment. **Every PR gets a verdict: `FAST_TRACK`, `REVIEW_REQUIRED`, or `RECOMMEND_CLOSE`.** Tested on [OpenClaw](https://github.com/openclaw/openclaw) (3,368 open PRs): cut the maintainer review queue by 36% and found 6% duplicate PRs in 30 seconds. [See the full report.](https://gist.github.com/pranayom) --- ## Installation ```bash # Core toolkit pip install oss-maintainer-toolkit # With PR triage / gatekeeper pipeline pip install "oss-maintainer-toolkit[gatekeeper]" # For development pip install -e ".[dev,gatekeeper]" ``` ### CLI usage ```bash maintainer assess --owner openclaw --repo openclaw --pr 18675 # PR triage ``` ### MCP server ```bash python -m oss_maintainer_toolkit.mcp # start the MCP server ``` --- ## Quick Start (GitHub Action) Copy this workflow into `.github/workflows/pr-triage.yml` in your repo: ```yaml name: PR Triage on: pull_request: types: [opened, synchronize, reopened] permissions: pull-requests: write contents: read jobs: triage: runs-on: ubuntu-latest steps: - uses: actions/checkout@v4 - uses: pranayom/oss-maintainer-toolkit@v1 with: github_token: ${{ secrets.GITHUB_TOKEN }} ``` That's it. Every new PR gets a scorecard comment with a verdict and flags. --- ## How It Works ``` PR opened | v [Tier 1: Embedding Dedup] — sentence-transformers, cosine similarity | Duplicates -> RECOMMEND_CLOSE (stop) v [Tier 2: Heuristic Scoring] — 7 deterministic rules, weighted scoring | Flagged -> REVIEW_REQUIRED (stop) v [Tier 3: Vision Alignment] — LLM compares PR against Vision Document (optional) | v FAST_TRACK ``` Tiers run strictly in sequence. Each tier is a gate — failures don't proceed to the next tier. This reserves LLM time for the minority of PRs where semantic judgment is actually useful. ### Tier 1 — Embedding Dedup (free, local) Computes semantic embeddings for PR title + description + diff using `all-MiniLM-L6-v2`. Flags duplicates above a cosine similarity threshold (default: 0.90). ### Tier 2 — Heuristic Scoring (free, deterministic) Seven rules scored against PR metadata: | Rule | What it catches | |---|---| | `new_account` | GitHub account < 90 days old | | `first_contribution` | No previously merged PRs on this repo | | `sensitive_paths` | Changes to auth, credentials, CI/CD, extensions | | `low_test_ratio` | Code added without proportional tests | | `unjustified_deps` | Dependency changes without explanation | | `large_diff_hiding` | Large PR with small sensitive changes buried in bulk | | `temporal_clustering` | Multiple new-account PRs within a short window | ### Tier 3 — Vision Alignment (optional, $0 via OpenRouter) Compares the PR diff against your project's Vision Document (a YAML file defining principles, anti-patterns, and focus areas). Uses OpenRouter free models. Requires an `OPENROUTER_API_KEY` (free at [openrouter.ai/keys](https://openrouter.ai/keys)). --- ## Inputs | Input | Required | Default | Description | |---|---|---|---| | `github_token` | Yes | — | GitHub token for API access (usually `secrets.GITHUB_TOKEN`) | | `vision_document` | No | `""` | Path to YAML vision document (relative to repo root) | | `openrouter_api_key` | No | `""` | OpenRouter API key for Tier 3 ($0 cost). Tier 3 skipped if not provided. | | `openrouter_model` | No | `openai/gpt-oss-120b:free` | OpenRouter model for Tier 3 | | `duplicate_threshold` | No | `0.9` | Cosine similarity threshold for duplicate detection | | `suspicion_threshold` | No | `0.6` | Suspicion score threshold for flagging | | `enforce_vision` | No | `false` | Enable Tier 3 vision alignment (set to `true` after reviewing your vision doc) | | `post_comment` | No | `true` | Post scorecard as a PR comment | ## Outputs | Output | Description | |---|---| | `verdict` | `FAST_TRACK`, `REVIEW_REQUIRED`, or `RECOMMEND_CLOSE` | | `scorecard_json` | Full scorecard as JSON for downstream CI steps | --- ## Vision Documents A Vision Document is an optional YAML file that defines what your project is trying to be. It enables Tier 3, where an LLM evaluates whether a PR aligns with your project's direction. Example structure: ```yaml project: my-project principles: - name: "Security First" description: "All changes touching auth or credentials require security review" - name: "Test Everything" description: "Every feature PR must include tests" anti_patterns: - "Adding dependencies without justification" - "Modifying CI/CD without maintainer approval" focus_areas: - "src/auth/" - "src/credentials/" - ".github/" ``` Place it at `.github/vision.yaml` and set `vision_document: ".github/vision.yaml"` in the action inputs. --- ## Example Scorecard Comment When the action runs on a PR, it posts a comment like: > ## &#x26A0; PR Triage: **REVIEW REQUIRED** > > > First-time contributor modifying sensitive paths without tests. > > | Dimension | Score | Summary | > |---|---|---| > | Hygiene & Dedup | `++++++++--` 0.80 | No duplicates found | > | Contributor Risk | `++++------` 0.40 | New account + sensitive paths | > > ### Flags > - [**HIGH**] **Sensitive Paths**: PR modifies `src/auth/oauth.ts`, `src/credentials/store.ts` > - [MEDIUM] **First Contribution**: No previously merged PRs from this author > - [MEDIUM] **Low Test Ratio**: 245 lines added, 0 test lines --- ## Roadmap - **PR Triage** — Shipped (v0.3.0) - **Issue Triage** — Dedup and classify issues - **Issue-to-PR Linking** — Suggest which PRs address which issues - **Label Automation** — Auto-classify PRs/issues into project label taxonomies - **Contributor Profiles** — Track contribution patterns and reliability - **Review Routing** — Suggest reviewers based on file ownership - **Smart Stale Detection** — Semantic staleness (superseded, merged elsewhere, blocked) - **Cross-PR Conflict Detection** — Surface PRs with overlapping file changes --- ## Evidence: OpenClaw Triage We ran this tool against 100 of OpenClaw's 3,368 open PRs: | Verdict | Count | Meaning | |---|---|---| | FAST_TRACK | 64 (64%) | Safe for quick review | | REVIEW_REQUIRED | 30 (30%) | Flagged — needs human attention | | RECOMMEND_CLOSE | 6 (6%) | Likely duplicate | - Found 3 duplicate clusters (6 PRs) at 0.90 threshold - 89% of PRs from first-time contributors - 40% touch security-sensitive paths - Extrapolated: ~200 closable duplicates in the full backlog --- ## Cost $0. All tiers run for free: - Tier 1: `sentence-transformers` on CPU (GitHub Actions runner) - Tier 2: Pure Python rules - Tier 3: OpenRouter free models (optional, free API key) --- ## License MIT
text/markdown
Pranay Om
null
null
null
null
dedup, gatekeeper, github, governance, issue-triage, maintainer, oss, pr-triage
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Software Development :: Quality Assurance" ]
[]
null
null
>=3.11
[]
[]
[]
[ "httpx>=0.25", "mcp[cli]>=1.2.0", "pydantic-settings>=2.0", "pydantic>=2.0", "rich>=13.0", "typer>=0.9", "build>=1.0; extra == \"dev\"", "pytest-asyncio>=0.23; extra == \"dev\"", "pytest>=7.0; extra == \"dev\"", "respx>=0.20; extra == \"dev\"", "twine>=5.0; extra == \"dev\"", "numpy>=1.24; extra == \"gatekeeper\"", "pyyaml>=6.0; extra == \"gatekeeper\"", "sentence-transformers>=2.0; extra == \"gatekeeper\"" ]
[]
[]
[]
[ "Homepage, https://github.com/pranayom/oss-maintainer-toolkit", "Repository, https://github.com/pranayom/oss-maintainer-toolkit", "Issues, https://github.com/pranayom/oss-maintainer-toolkit/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:34:45.380824
oss_maintainer_toolkit-0.4.0.tar.gz
126,290
6e/a1/0402082dcf6f76a6c1709687fc25b71f0677f5ef921487fdf5b7c5f19f85/oss_maintainer_toolkit-0.4.0.tar.gz
source
sdist
null
false
018ed520e0d15566f12416fc4e9b89c8
f4178cb98b883f19c40bad77a06547201a68c5b19592301861001edc3f0ddca6
6ea10402082dcf6f76a6c1709687fc25b71f0677f5ef921487fdf5b7c5f19f85
MIT
[ "LICENSE" ]
208
2.4
pybiolib
1.3.216
BioLib Python Client
# PyBioLib PyBioLib is a Python package for running BioLib applications from Python scripts and the command line. ### Python Example ```python # pip3 install -U pybiolib import biolib samtools = biolib.load('samtools/samtools') print(samtools.cli(args='--help')) ``` ### Command Line Example ```bash pip3 install -U pybiolib biolib run samtools/samtools --help ```
text/markdown
null
biolib <hello@biolib.com>
null
null
null
biolib
[ "Operating System :: OS Independent", "Programming Language :: Python :: 3" ]
[]
null
null
>=3.6.3
[]
[]
[]
[ "click>=8.0.0", "importlib-metadata>=1.6.1", "typing-extensions>=4.1.0; python_version < \"3.11\"", "docker>=5.0.3; extra == \"compute-node\"", "flask>=2.0.1; extra == \"compute-node\"", "gunicorn>=20.1.0; extra == \"compute-node\"", "docker>=5.0.3; extra == \"sdk\"", "pyyaml>=5.3.1; extra == \"sdk\"", "rich>=12.4.4; extra == \"sdk\"" ]
[]
[]
[]
[ "Homepage, https://github.com/biolib" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:34:44.540403
pybiolib-1.3.216.tar.gz
144,161
fa/e8/2f4761df5f360445b75e671b9e0f7e1ba536e60195b9235035ff98d8c65e/pybiolib-1.3.216.tar.gz
source
sdist
null
false
962752cf37fdd52affea070e58d68d13
519f5811fb70813fbceaea96e3682cd1535a7720ba2c5b56ccca1b776e05a473
fae82f4761df5f360445b75e671b9e0f7e1ba536e60195b9235035ff98d8c65e
MIT
[ "LICENSE" ]
283
2.4
snailz
5.5.4
Synthetic data generator for snail mutation survey
# Snailz <img src="https://raw.githubusercontent.com/gvwilson/snailz/refs/heads/main/pages/img/snailz-logo.svg" alt="snail logo" width="200px"> `snailz` is a synthetic data generator that models a study of snails in the Pacific Northwest which are growing to unusual size as a result of exposure to pollution. The package generates fully-reproducible datasets of varying sizes and with varying statistical properties, and is intended for classroom use. For example, an instructor can give each learner a unique dataset to analyze, while learners can test their analysis pipelines using datasets they generate themselves. > *The Story* > > Years ago, > logging companies dumped toxic waste in a remote region of Vancouver Island. > As the containers leaked and the pollution spread, > some snails in the region began growing unusually large. > Your team is now collecting and analyzing specimens from affected regions > to determine if exposure to pollution is responsible. ## Usage: ``` usage: snailz [-h] [--defaults] [--outdir OUTDIR] [--override OVERRIDE [OVERRIDE ...]] [--params PARAMS] [--profile] options: -h, --help show this help message and exit --defaults show default parameters as JSON --outdir OUTDIR output directory --override OVERRIDE [OVERRIDE ...] name=value parameters to override defaults --params PARAMS specify JSON parameter file --profile enable profiling ``` See the documentation of the `Parameters` class for a description of data generation parameters. ## Schema <img src="https://raw.githubusercontent.com/gvwilson/snailz/refs/heads/main/pages/img/schema.svg" alt="snailz schema"> An asterisk beside the name of a field indicates that the value may be missing (i.e., the field may be `NULL` in the final database). | table | field | type | purpose | | -------------- | ------------- | ----- | ------- | | grid | ident | text | unique identifier for each survey grid | | | size | int | height and width of survey grid in cells | | | spacing | float | size of survey grid cell (meters) | | | lat0 | float | southernmost latitude of grid (fractional degrees) | | | lon0 | float | westernmost longitude of grid (fractional degrees) | | | | | | | grid_cells | grid_id | text | foreign key reference to grid | | | lat | float | foreign key reference to grid cell | | | lon | float | foreign key reference to grid cell | | | value | float | pollution measurement in that grid cell | | | | | | | machine | ident | text | unique identifier for each piece of laboratory equipment | | | name | text | name of piece of laboratory equipment | | | | | | | person | ident | text | unique identifier for member of staff | | | family | text | family name of staff member | | | personal | text | personal name of staff member | | | supervisor_id | text* | foreign key reference to person's supervisor | | | | | | | rating | person_id | text | foreign key reference to person | | | machine_id | text | foreign key reference to machine | | | certified | bool | whether person is certified to use machine | | | | | | | assay | ident | text | unique identifier for soil assay | | | lat | float | foreign key reference to grid cell | | | lon | float | foreign key reference to grid cell | | | person_id | text | foreign key reference to person who did assay | | | machine_id | text | foreign key reference to machine used to do assay | | | performed* | date | date that assay was done | | | | | | | assay_readings | assay_id | text | foreign key reference to assay | | | reading_id | int | serial number within assay | | | contents | text | "C" or "T" showing control or treatment | | | reading | float | pollution measurement | | | | | | | species | reference | text | reference genome | | | susc_locus | int | location of susceptible locus within genome | | | susc_base | text | base that causes significant mutation at that locus | | | | | | | species_loci | ident | int | unique locus serial number | | | locus | int | locus where mutation might occur | | | | | | | specimen | ident | text | unique identifier for specimen | | | lat | float | foreign key reference to grid cell | | | lon | float | foreign key reference to grid cell | | | genome | text | specimen genome | | | mass | float | specimen mass (g) | | | diameter | float | specimen diameter (mm) | | | collected* | date | when specimen was collected | ## Colophon `snailz` was inspired by the [Palmer Penguins][penguins] dataset and by conversations with [Rohan Alexander][alexander-rohan] about his book [*Telling Stories with Data*][telling-stories]. My thanks to everyone who built the tools this project relies on, including: - [faker][faker] for synthesizing data. - [mkdocs][mkdocs] for documentation. - [ruff][ruff] for checking the code. - [sqlite][sqlite] and [sqlite-utils][sqlite-utils] for persistence. - [taskipy][taskipy] for running tasks. - [uv][uv] for managing packages and the virtual environment. The snail logo was created by [sunar.ko][snail-logo]. ## Acknowledgments - [*Greg Wilson*][wilson-greg] is a programmer, author, and educator based in Toronto. He was the co-founder and first Executive Director of Software Carpentry and received ACM SIGSOFT's Influential Educator Award in 2020. [alexander-rohan]: https://rohanalexander.com/ [faker]: https://faker.readthedocs.io/ [mkdocs]: https://www.mkdocs.org/ [penguins]: https://allisonhorst.github.io/palmerpenguins/ [ruff]: https://docs.astral.sh/ruff/ [snail-logo]: https://www.vecteezy.com/vector-art/7319786-snails-logo-vector-on-white-background [sqlite]: https://sqlite.org/ [sqlite-utils]: https://sqlite-utils.datasette.io/en/stable/ [taskipy]: https://pypi.org/project/taskipy/ [telling-stories]: https://tellingstorieswithdata.com/ [uv]: https://docs.astral.sh/uv/ [wilson-greg]: https://third-bit.com/
text/markdown
null
Greg Wilson <gvwilson@third-bit.com>
null
null
null
open science, synthetic data
[]
[]
null
null
>=3.12
[]
[]
[]
[ "faker>=40.1.2", "numpy>=2.4.1", "pillow>=12.1.0", "sqlite-utils==4.0a1", "twine>=6.2.0", "ty>=0.0.14" ]
[]
[]
[]
[ "Repository, https://github.com/gvwilson/snailz", "Documentation, https://snailz.readthedocs.io" ]
twine/6.2.0 CPython/3.12.12
2026-02-20T21:34:23.506447
snailz-5.5.4.tar.gz
828,843
c5/ed/e3d8536f7d668058c7536153582b72221d7dbdbaa02536c7d9a0fe9e8d13/snailz-5.5.4.tar.gz
source
sdist
null
false
61415615e78f9c7b0909e01216f36e3d
0a9ef62f98bd484cd1fd1b6e259ea5f0d3d994b7dbb39d035a635e131bdb90f2
c5ede3d8536f7d668058c7536153582b72221d7dbdbaa02536c7d9a0fe9e8d13
null
[ "LICENSE.md" ]
193
2.4
lightningrod-ai
0.1.14
Python SDK for Lightning Rod AI-powered forecasting dataset generation
<div align="center"> <!-- Note: only an absolute image URL works on PyPi: https://pypi.org/project/lightningrod-ai --> <img src="https://github.com/lightning-rod-labs/lightningrod-python-sdk/blob/main/banner.png?raw=true" alt="Lightning Rod Labs" /> </div> # Lightning Rod Python SDK [![Beta](https://img.shields.io/badge/beta-0.1.14-orange)](https://pypi.org/project/lightningrod-ai/0.1.14/) The Lightning Rod SDK provides a simple Python API for generating custom forecasting datasets to train your LLMs. Transform news articles, documents, and other real-world data into high-quality training samples automatically. Based on our research: [Future-as-Label: Scalable Supervision from Real-World Outcomes](https://arxiv.org/abs/2601.06336) ## 👋 Quick Start ### 1. Install the SDK ```bash pip install lightningrod-ai ``` ### 2. Get your API key Sign up at [dashboard.lightningrod.ai](https://dashboard.lightningrod.ai/?redirect=/api) to get your API key and **$50 of free credits**. ### 3. Generate your first dataset Generate **1000+ forecasting questions in minutes** - from raw sources to labeled dataset, automatically. ⚡ ```python from lightningrod import LightningRod, BinaryAnswerType, QuestionPipeline, NewsSeedGenerator, ForwardLookingQuestionGenerator, WebSearchLabeler lr = LightningRod(api_key="your-api-key") binary_answer = BinaryAnswerType() pipeline = QuestionPipeline( seed_generator=NewsSeedGenerator( start_date=datetime.now() - timedelta(days=90), end_date=datetime.now(), search_query=["Trump"], ), question_generator=ForwardLookingQuestionGenerator( instructions="Generate binary forecasting questions about Trump's actions and decisions.", examples=[ "Will Trump impose 25% tariffs on all goods from Canada by February 1, 2025?", "Will Pete Hegseth be confirmed as Secretary of Defense by February 15, 2025?", ] ), labeler=WebSearchLabeler(answer_type=binary_answer), ) dataset = lr.transforms.run(pipeline, max_questions=3000) dataset.flattened() # Ready-to-use data for your training pipelines ``` **We use this to generate the [Future-as-Label training dataset](https://huggingface.co/datasets/LightningRodLabs/future-as-label-paper-training-dataset) for our research paper.** ## ✨ Examples We have some example notebooks to help you get started! If you have trouble using the SDK, please submit an issue on Github. | Example Name | Path | Google Colab Link | |--------------|------|-------------------| | Quick Start | `notebooks/01_quick_start.ipynb` | [![Open in Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/lightning-rod-labs/lightningrod-python-sdk/blob/main/notebooks/01_quick_start.ipynb) | | News Datasource | `notebooks/02_news_datasource.ipynb` | [![Open in Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/lightning-rod-labs/lightningrod-python-sdk/blob/main/notebooks/02_news_datasource.ipynb) | | Custom Documents | `notebooks/03_custom_documents_datasource.ipynb` | [![Open in Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/lightning-rod-labs/lightningrod-python-sdk/blob/main/notebooks/03_custom_documents_datasource.ipynb) | | Binary Answer Type | `notebooks/04_binary_answer_type.ipynb` | [![Open in Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/lightning-rod-labs/lightningrod-python-sdk/blob/main/notebooks/04_binary_answer_type.ipynb) | | Continuous Answer Type | `notebooks/05_continuous_answer_type.ipynb` | [![Open in Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/lightning-rod-labs/lightningrod-python-sdk/blob/main/notebooks/05_continuous_answer_type.ipynb) | | Multiple Choice Answer Type | `notebooks/06_multiple_choice_answer_type.ipynb` | [![Open in Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/lightning-rod-labs/lightningrod-python-sdk/blob/main/notebooks/06_multiple_choice_answer_type.ipynb) | | Free Response Answer Type | `notebooks/07_free_response_answer_type.ipynb` | [![Open in Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/lightning-rod-labs/lightningrod-python-sdk/blob/main/notebooks/07_free_response_answer_type.ipynb) | For complete API reference documentation, see [API.md](API.md). This includes overview of the core system concepts, methods and types.
text/markdown
null
Lightning Rod Labs <support@lightningrod.ai>
null
null
MIT License Copyright (c) 2025 Lightning Rod Labs Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
null
[]
[]
null
null
>=3.10
[]
[]
[]
[ "requests>=2.31.0", "pydantic>=2.0.0", "httpx>=0.25.0", "attrs>=23.1.0", "python-dateutil>=2.8.0", "pyarrow>=14.0.0", "fsspec>=2023.0.0", "rich>=13.0.0", "openapi-python-client>=0.15.0; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://lightningrod.ai/sdk", "Repository, https://github.com/lightning-rod-labs/lightningrod-python-sdk" ]
twine/6.2.0 CPython/3.10.18
2026-02-20T21:33:12.730831
lightningrod_ai-0.1.14.tar.gz
81,143
d7/e3/990ba65ed8835a3e6376e7281b951a235fe3e2aa24ed1d5224baac8642be/lightningrod_ai-0.1.14.tar.gz
source
sdist
null
false
d50128843f785c69c51fdf6e79c0a596
3a3fccea0baf36e7bec930c090598596bdf8a2d4b1290dee1277c15eb8ecf568
d7e3990ba65ed8835a3e6376e7281b951a235fe3e2aa24ed1d5224baac8642be
null
[ "LICENSE" ]
225
2.4
finanfut-billing-sdk
2.1.13
Python SDK for Finanfut Billing External API
# Finanfut Billing Python SDK Client oficial sincrònic per consumir la **Finanfut Billing External API (`/external/v1`)** amb models compatibles amb **Pydantic v2**, ara preparat per treballar amb **Business Units**. ## Instal·lació - Pydantic 2.x: `pip install finanfut-billing-sdk>=2.0` - Pydantic 1.x: `pip install finanfut-billing-sdk<2.0` - Des del repositori local: `pip install -e backend/sdk` Dependències principals: - `pydantic>=2.0,<3.0` - `requests>=2.31` ## Configuració bàsica i Business Units ```python from finanfut_billing_sdk import FinanfutBillingClient client = FinanfutBillingClient( base_url="https://api.finanfut-billing.com", api_key="sk_live_xxx", business_unit_id="bu_default", # opcional: aplicada a serveis/factures/liquidacions per defecte ) ``` ### Com funciona `business_unit_id` - **Global al client:** passa `business_unit_id` al constructor i s'aplicarà automàticament a les operacions compatibles. - **Per operació:** pots sobreescriure-la en cada mètode (`business_unit_id="bu_alt"`). - **Endpoints sense BU:** tax rates i partner payment methods són d'abast de companyia i ignoren la BU (l'SDK emet un avís si n'hi ha una definida). - **Liquidacions (settlements):** la BU és opcional però s'envia quan està disponible per enrutar payouts. - **External API:** serveis i factures externes accepten BU però actualment poden ignorar-la; l'SDK ja no la tracta com a obligatòria. ## Exemples d'ús ### Crear producte/servei amb BU ```python from decimal import Decimal from finanfut_billing_sdk.models import ExternalServiceUpsertRequest payload = ExternalServiceUpsertRequest( external_reference="service_abc", type="service", name="Monthly subscription", description="Access to premium content", price=Decimal("29.90"), vat_rate_code="vat_21", ) service = client.upsert_service(payload) # usa la BU global ``` ### Crear factura amb BU (sobre-escrivint la BU global) ```python from finanfut_billing_sdk.models import ExternalInvoiceCreateRequest, ExternalInvoiceLine invoice = client.create_invoice( ExternalInvoiceCreateRequest( client_external_reference="client_123", currency="EUR", lines=[ ExternalInvoiceLine( service_external_reference="service_abc", description="Premium plan", qty=1, price=29.90, vat_rate_id="tax_rate_uuid", ), ], ), business_unit_id="bu_sales", # prioritat respecte la BU global ) ``` ### Operacions sense BU (àmbit de companyia) ```python # Els tax rates i partner payment methods ignoren la BU. client.list_tax_rates() client.partner_payment_methods.list_partner_payment_methods() ``` ### Idempotència en liquidacions ```python settlement = client.settlements.create_settlement( payload, idempotency_key="settlement-create-2024-12-31", ) ``` ### Enviar factura i registrar pagament ```python from finanfut_billing_sdk.models import ExternalInvoiceEmailRequest, ExternalPaymentCreateRequest email = client.send_invoice_email( invoice.invoice_id, ExternalInvoiceEmailRequest(subject="La teva factura", body="Adjunt trobaràs el PDF"), ) payment = client.register_payment( invoice.invoice_id, ExternalPaymentCreateRequest(amount=29.90, method="stripe"), ) ``` ### Checkout i onboarding de Stripe Connect ```python from finanfut_billing_sdk.models import ExternalCheckoutCreateRequest, ExternalConnectOnboardRequest checkout = client.payments.create_checkout( "stripe", ExternalCheckoutCreateRequest( amount=29.90, currency="EUR", business_unit_id="bu_sales", provider_payload={"payment_method_types": ["card"]}, ), ) connect = client.payments.connect_onboard( "stripe", ExternalConnectOnboardRequest( provider_id="provider_uuid", return_url="https://app.example.com/connect/return", refresh_url="https://app.example.com/connect/refresh", ), ) ``` ### Checkout sessions de Stripe (external) ```python from finanfut_billing_sdk.models import ExternalCheckoutSessionCreateRequest session = client.payments.create_checkout_session( ExternalCheckoutSessionCreateRequest( amount=49.90, currency="EUR", success_url="https://app.example.com/ok", cancel_url="https://app.example.com/cancel", description="Pagament", ), idempotency_key="checkout-session-2024-12-01", ) ``` ### Subscripcions BU ```python from finanfut_billing_sdk.models import SubscriptionPricingSnapshot, SubscriptionStartRequest payload = SubscriptionStartRequest( request_id="sports-pro-2025-01", business_unit_id="bu_sales", subject_type="team", subject_id="team_123", billing_client_id="client_uuid", bu_plan_ref="pro_v3", pricing_snapshot=SubscriptionPricingSnapshot( amount=29.9, currency="EUR", interval="month", ), success_url="https://app.example.com/billing/success", cancel_url="https://app.example.com/billing/cancel", ) response = client.subscriptions.start_subscription(payload) ``` ## Errors ```python from finanfut_billing_sdk.errors import ( FinanfutBillingAuthError, FinanfutBillingServiceError, FinanfutBillingValidationError, ) try: client.list_tax_rates() except FinanfutBillingAuthError: print("API key incorrecta o sense permisos") except FinanfutBillingValidationError as e: print("Error de validació:", e.payload) except FinanfutBillingServiceError as e: print(f"Error de servei ({e.request_id}): {e.error}") ``` Els errors del backend inclouen sempre `error`, `message` i `request_id`. ## Publicació a PyPI El paquet està preparat per publicar-se a PyPI quan es creen tags `v*` al repositori. El workflow `publish-sdk.yml` valida la versió (`__version__`) i fa l'upload amb Twine.
text/markdown
Finanfut
null
null
null
MIT License Copyright (c) 2025 erovirafinanfut Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
null
[]
[]
null
null
>=3.9
[]
[]
[]
[ "pydantic<3.0,>=2.0", "requests>=2.31" ]
[]
[]
[]
[ "Homepage, https://finanfut.com", "Source, https://github.com/finanfut/finanfut-billing" ]
twine/6.2.0 CPython/3.14.0
2026-02-20T21:32:56.399954
finanfut_billing_sdk-2.1.13.tar.gz
14,309
68/a7/2d0ca1501c786014d559d929896c84bcd35fcebcc47dafa050db0a9f22a0/finanfut_billing_sdk-2.1.13.tar.gz
source
sdist
null
false
3bd15d5355aeea87f9b438bb27935999
d1284b7980db5e20dfbfc78ac792c42db45e803dbd34c961e5b0d9975717ce29
68a72d0ca1501c786014d559d929896c84bcd35fcebcc47dafa050db0a9f22a0
null
[ "LICENSE" ]
212
2.4
linden
0.5.1
A Python framework for building AI agents with multi-provider LLM support, persistent memory, and function calling capabilities.
# Linden <div align="center"> <img src="https://raw.githubusercontent.com/matstech/linden/main/doc/logo.png" alt="Linden Logo" width="200"/> </div> <div align="center"> <p><em>A Python framework for building AI agents with multi-provider LLM support, persistent memory, and function calling capabilities.</em></p> </div> <div align="center"> ![Python](https://img.shields.io/badge/Python-3776AB?style=flat&logo=python&logoColor=white) ![License](https://img.shields.io/github/license/matstech/linden) ![Version](https://img.shields.io/github/v/release/matstech/linden) ![CI](https://github.com/matstech/linden/actions/workflows/python-publish.yml/badge.svg) ![Coverage](https://codecov.io/gh/matstech/linden/branch/main/graph/badge.svg) <!--![Contributors](https://img.shields.io/github/contributors/matstech/linden)--> </div> ## Table of Contents - [Overview](#overview) - [Features](#features) - [Installation](#installation) - [Requirements](#requirements) - [Quick Start](#quick-start) - [Agent Configuration](#agent-configuration) - [Basic Agent Setup](#basic-agent-setup) - [Agent with Function Calling](#agent-with-function-calling) - [Streaming Responses](#streaming-responses) - [Structured Output with Pydantic](#structured-output-with-pydantic) - [Configuration](#configuration) - [Environment Variables](#environment-variables) - [Architecture](#architecture) - [Core Components](#core-components) - [Memory Architecture](#memory-architecture) - [Function Tool Definition](#function-tool-definition) - [Advanced Usage](#advanced-usage) - [Multi-Turn Conversations](#multi-turn-conversations) - [Error Handling and Retries](#error-handling-and-retries) - [Memory Management](#memory-management) - [Provider-Specific Features](#provider-specific-features) - [API Reference](#api-reference) - [AgentRunner](#agentrunner) - [Memory Classes](#memory-classes) - [Configuration](#configuration-1) - [Error Types](#error-types) - [Contributing](#contributing) - [License](#license) - [Support](#support) ## Overview Linden is a comprehensive AI agent framework that provides a unified interface for interacting with multiple Large Language Model (LLM) providers including OpenAI, Anthropic, Groq, Google (Gemini), and Ollama. It features persistent conversation memory, automatic tool/function calling, and robust error handling for building production-ready AI applications. ## Features - **Multi-Provider LLM Support**: Seamless integration with OpenAI, Anthropic, Groq, Google (Gemini), and Ollama - **Persistent Memory**: Long-term conversation memory using FAISS vector storage and embeddings - **Function Calling**: Automatic parsing and execution of tools with Google-style docstring support - **Streaming Support**: Real-time response streaming for interactive applications - **Thread-Safe Memory**: Concurrent agent support with isolated memory per agent - **Configuration Management**: Flexible TOML-based configuration with environment variable support - **Type Safety**: Full Pydantic model support for structured outputs and agent configuration - **Error Handling**: Comprehensive error handling with retry mechanisms - **Validated Configuration**: Strict parameter validation with Pydantic's AgentConfiguration model ## Installation ```bash pip install linden ``` ## Requirements - Python >= 3.9 - Dependencies automatically installed: - `openai` - OpenAI API client - `anthropic` - Anthropic API client - `groq` - Groq API client - `google-genai` - Google Gemini client - `ollama` - Ollama local LLM client - `pydantic` - Data validation and serialization - `mem0` - Memory management - `docstring_parser` - Function documentation parsing ## Agent Configuration Linden uses a Pydantic model called `AgentConfiguration` to define and validate all agent parameters. This provides: - Strong typing and validation for all agent parameters - Rejection of invalid or unsupported parameters - Default values for optional parameters - Clear documentation of configuration options Example of using `AgentConfiguration`: ```python from linden.core import AgentConfiguration, Provider config = AgentConfiguration( user_id="user123", name="assistant", model="gpt-4", temperature=0.7, system_prompt="You are a helpful AI assistant.", tools=[get_weather], # Optional list of callable functions output_type=PersonInfo, # Optional Pydantic model for structured output client=Provider.OPENAI, # AI provider enum retries=3 # Retry attempts for failed requests ) # Create agent with configuration agent = AgentRunner(config=config) ``` ## Quick Start ### Basic Agent Setup ```python from linden.core import AgentRunner, AgentConfiguration, Provider # Create an agent configuration config = AgentConfiguration( user_id="user123", name="assistant", model="gpt-4", temperature=0.7, system_prompt="You are a helpful AI assistant.", client=Provider.OPENAI ) # Initialize the agent with configuration agent = AgentRunner(config=config) # Ask a question response = agent.run("What is the capital of France?") print(response) ``` ### Agent with Function Calling ```python def get_weather(location: str, units: str = "celsius") -> str: """Get current weather for a location. Args: location (str): The city name or location units (str, optional): Temperature units (celsius/fahrenheit). Defaults to celsius. Returns: str: Weather information """ return f"The weather in {location} is 22°{units[0].upper()}" # Create agent configuration with tools config = AgentConfiguration( user_id="user123", name="weather_bot", model="gpt-4", temperature=0.7, system_prompt="You are a weather assistant.", tools=[get_weather], client=Provider.OPENAI ) # Initialize the agent agent = AgentRunner(config=config) response = agent.run("What's the weather in Paris?") print(response) ``` ### Streaming Responses ```python # Stream responses for real-time interaction for chunk in agent.run("Tell me a story", stream=True): print(chunk, end="", flush=True) ``` ### Structured Output with Pydantic ```python from pydantic import BaseModel from linden.core import AgentRunner, AgentConfiguration, Provider class PersonInfo(BaseModel): name: str age: int occupation: str # Create agent configuration with output_type for structured outputs config = AgentConfiguration( user_id="user123", name="extractor", model="gpt-4", temperature=0.1, system_prompt="Extract person information from text.", output_type=PersonInfo, client=Provider.OPENAI ) # Initialize the agent agent = AgentRunner(config=config) result = agent.run("John Smith is a 30-year-old software engineer.") print(f"Name: {result.name}, Age: {result.age}") ``` ## Configuration Create a `config.toml` file in your project root: ```toml [models] dec = "gpt-4" tool = "gpt-4" extractor = "gpt-3.5-turbo" speaker = "gpt-4" [openai] api_key = "your-openai-api-key" timeout = 30 [anthropic] api_key = "your-anthropic-api-key" timeout = 30 max_tokens = 1024 #example [groq] base_url = "https://api.groq.com/openai/v1" api_key = "your-groq-api-key" timeout = 30 [ollama] timeout = 60 [google] api_key = "your-google-api-key" timeout = 60 [memory] path = "./memory_db" collection_name = "agent_memories" ``` ### Environment Variables Set your API keys as environment variables: ```bash export OPENAI_API_KEY="your-openai-api-key" export ANTHROPIC_API_KEY="your-anthropic-api-key" export GROQ_API_KEY="your-groq-api-key" export GOOGLE_API_KEY="your-google-api-key" ``` ## Architecture ### Core Components #### AgentRunner The main agent orchestrator that handles: - LLM interaction and response processing - Tool calling and execution - Memory management - Error handling and retries - Streaming and non-streaming responses #### Memory System - **AgentMemory**: Per-agent conversation history and semantic search - **MemoryManager**: Thread-safe singleton for shared vector storage - **Persistent Storage**: FAISS-based vector database for long-term memory #### AI Clients Abstract interface with concrete implementations: - **OpenAiClient**: OpenAI GPT models - **AnthropicClient**: Anthropic Claude models - **GroqClient**: Groq inference API - **GoogleClient**: Google Gemini models - **Ollama**: Local LLM execution #### Function Calling - Automatic parsing of Google-style docstrings - JSON Schema generation for tool descriptions - Type-safe argument parsing and validation - Error handling for tool execution ### Memory Architecture The memory system uses a shared FAISS vector store with agent isolation: ```python # Each agent has isolated memory agent1 = AgentRunner(name="agent1", ...) agent2 = AgentRunner(name="agent2", ...) # Memories are automatically isolated by agent_id agent1.run("Remember I like coffee") agent2.run("Remember I like tea") # Each agent only retrieves its own memories ``` ### Function Tool Definition Functions must use Google-style docstrings for automatic parsing: ```python def search_database(query: str, limit: int = 10, filters: dict = None) -> list: """Search the knowledge database. Args: query (str): The search query string limit (int, optional): Maximum results to return. Defaults to 10. filters (dict, optional): Additional search filters: category (str): Filter by category date_range (str): Date range in ISO format Returns: list: List of search results with metadata """ # Implementation here pass ``` ## Advanced Usage ### Multi-Turn Conversations ```python from linden.core import AgentRunner, AgentConfiguration # Create agent configuration config = AgentConfiguration( user_id="user123", name="chat_bot", model="gpt-4", temperature=0.7, system_prompt="You are a helpful assistant." ) agent = AgentRunner(config=config) # Conversation maintains context automatically agent.run("My name is Alice") agent.run("What's my name?") # Will remember "Alice" agent.run("Tell me about my previous question") # Has full context ``` ### Error Handling and Retries ```python from linden.core import AgentRunner, AgentConfiguration from linden.core.model import ToolError, ToolNotFound # Configure agent with retries config = AgentConfiguration( user_id="user123", name="robust_agent", model="gpt-4", temperature=0.7, system_prompt="You are a helpful assistant.", retries=3 # Retry failed calls up to 3 times ) agent = AgentRunner(config=config) try: response = agent.run("Complex query that might fail") except ToolError as e: print(f"Tool execution failed: {e.message}") except ToolNotFound as e: print(f"Tool not found: {e.message}") ``` ### Memory Management ```python # Reset agent memory agent.reset() # Add context without user interaction agent.add_to_context("Important context information", persist=True) # Get conversation history history = agent.memory.get_conversation("Current query") ``` ### Provider-Specific Features ```python from linden.core import AgentRunner, AgentConfiguration, Provider # Use Anthropic Claude models claude_config = AgentConfiguration( user_id="user123", name="claude_agent", model="claude-3-opus-20240229", system_prompt="You are a helpful assistant.", temperature=0.7, client=Provider.ANTHROPIC ) claude_agent = AgentRunner(config=claude_config) # Use local Ollama models ollama_config = AgentConfiguration( user_id="user123", name="local_agent", model="llama2", system_prompt="You are a helpful assistant.", temperature=0.7, client=Provider.OLLAMA ) local_agent = AgentRunner(config=ollama_config) # Use Groq for fast inference groq_config = AgentConfiguration( user_id="user123", name="fast_agent", model="mixtral-8x7b-32768", system_prompt="You are a helpful assistant.", temperature=0.7, client=Provider.GROQ ) fast_agent = AgentRunner(config=groq_config) # Use Google Gemini # Assicurati di avere GOOGLE_API_KEY configurata o nel config.toml # tools devono essere nel formato supportato da Gemini (function_declarations) gemini_config = AgentConfiguration( user_id="user123", name="gemini_agent", model="gemini-1.5-flash", system_prompt="You are a helpful assistant.", temperature=0.7, client=Provider.GOOGLE, ) gemini_agent = AgentRunner(config=gemini_config) ``` ## API Reference ### AgentConfiguration #### Parameters - `user_id` (str): Unique identifier for the user - `name` (str): Unique agent identifier (defaults to UUID4) - `model` (str): LLM model name - `temperature` (float): Response randomness (0-1) - `system_prompt` (str): System instruction - `tools` (list[Callable], optional): Available functions (defaults to empty list) - `output_type` (BaseModel, optional): Structured output schema (defaults to None) - `client` (Provider): LLM provider selection (defaults to Provider.OLLAMA) - `retries` (int): Maximum retry attempts (defaults to 3) #### Features - Type validation for all parameters - Strict parameter validation (rejects unknown parameters) - Default values for optional parameters ### AgentRunner #### Constructor Parameters - `config` (AgentConfiguration): Configuration object for the agent with all the necessary settings #### Methods - `run(user_question: str, stream: bool = False)`: Execute agent query - `reset()`: Clear conversation history - `add_to_context(content: str, persist: bool = False)`: Add contextual information ### Memory Classes #### AgentMemory - `record(message: str, persist: bool = False)`: Store message - `get_conversation(user_input: str)`: Retrieve relevant context - `reset()`: Clear agent memory #### MemoryManager (Singleton) - `get_memory()`: Access shared memory instance - `get_all_agent_memories(agent_id: str = None)`: Retrieve stored memories ### Configuration #### ConfigManager - `initialize(config_path: str | Path)`: Load configuration file - `get(config_path: Optional[str | Path] = None)`: Get configuration instance - `reload()`: Refresh configuration from file ## Error Types - `ToolNotFound`: Requested function not available - `ToolError`: Function execution failed - `ValidationError`: Pydantic model validation failed - `RequestException`: HTTP/API communication error ## Contributing 1. Fork the repository 2. Create a feature branch (`git checkout -b feature/new-feature`) 3. Commit your changes (`git commit -am 'Add new feature'`) 4. Push to the branch (`git push origin feature/new-feature`) 5. Create a Pull Request ## License This project is licensed under the MIT License - see the LICENSE file for details. ## Support - GitHub Issues: [https://github.com/matstech/linden/issues](https://github.com/matstech/linden/issues) - Documentation: [https://github.com/matstech/linden](https://github.com/matstech/linden)
text/markdown
null
Matteo Stabile <matteo.stabile2@gmail.com>
null
null
null
null
[ "Programming Language :: Python :: 3", "Operating System :: OS Independent" ]
[]
null
null
>=3.10
[]
[]
[]
[ "docstring_parser==0.17.0", "pydantic==2.11.7", "Requests==2.32.5", "mem0ai==0.1.116", "anthropic==0.64.0", "ollama==0.5.3", "openai==1.101.0", "groq==0.31.0", "google-genai==1.63.0", "tomli", "pytest>=8.0.0; extra == \"test\"", "pytest-cov>=5.0.0; extra == \"test\"", "pytest-mock>=3.12.0; extra == \"test\"", "pytest-asyncio>=0.23.6; extra == \"test\"", "responses>=0.25.0; extra == \"test\"", "black>=24.1.0; extra == \"dev\"", "isort>=5.13.0; extra == \"dev\"", "flake8>=7.0.0; extra == \"dev\"", "mypy>=1.8.0; extra == \"dev\"", "pytest>=8.0.0; extra == \"all\"", "pytest-cov>=5.0.0; extra == \"all\"", "pytest-mock>=3.12.0; extra == \"all\"", "pytest-asyncio>=0.23.6; extra == \"all\"", "responses>=0.25.0; extra == \"all\"", "black>=24.1.0; extra == \"all\"", "isort>=5.13.0; extra == \"all\"", "flake8>=7.0.0; extra == \"all\"", "mypy>=1.8.0; extra == \"all\"" ]
[]
[]
[]
[ "Homepage, https://github.com/matstech/linden", "Bug Tracker, https://github.com/matstech/linden/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-20T21:32:55.582506
linden-0.5.1.tar.gz
28,977
45/23/1a77939ad120a40c05ad2324e648c2b46bfba8995e3bf6e4da29d389addc/linden-0.5.1.tar.gz
source
sdist
null
false
7dedf4e3ba0fcc1d407a7543fea0d581
4e258fd4c31fcf3c32fae1a9b81d62c5857270c86d34e66cea6da8075b8ded9c
45231a77939ad120a40c05ad2324e648c2b46bfba8995e3bf6e4da29d389addc
MIT
[ "LICENSE" ]
209
2.4
aicippy
1.6.12
Enterprise-grade multi-agent CLI system powered by AWS Bedrock
# AiCippy Enterprise-grade multi-agent CLI system powered by AWS Bedrock. **Copyright (c) 2024-2026 AiVibe Software Services Pvt Ltd. All rights reserved.** ISO 27001:2022 Certified | NVIDIA Inception Partner | AWS Activate | Microsoft for Startups ## Overview AiCippy is a production-grade, multi-agent command-line system that orchestrates up to 10 parallel AI agents to complete complex tasks. Built on AWS Bedrock Agents, it provides: - Multi-agent orchestration with parallel execution - Real-time WebSocket communication - MCP-style tool connectors for AWS, GitHub, Firebase, and more - Knowledge Base integration with automated feed ingestion - Rich terminal UI with live progress and agent status ## Installation ```bash pip install aicippy ``` ## Quick Start ```bash # Authenticate aicippy login # Initialize in a project aicippy init # Interactive mode aicippy # Single query aicippy chat "Explain this codebase" # Multi-agent task aicippy run "Deploy infrastructure to AWS" --agents 5 ``` ## Commands | Command | Description | |---------|-------------| | `aicippy` | Start interactive session | | `aicippy login` | Authenticate with Cognito | | `aicippy logout` | Clear credentials | | `aicippy init` | Initialize project context | | `aicippy chat <msg>` | Single query mode | | `aicippy run <task>` | Execute with agents | | `aicippy config` | Show/edit configuration | | `aicippy status` | Agent status | | `aicippy usage` | Token usage | | `aicippy upgrade` | Self-update | ## Interactive Commands | Command | Description | |---------|-------------| | `/help` | Show all commands | | `/model <name>` | Switch model (opus/sonnet/llama) | | `/mode <name>` | Change mode (agent/edit/research/code) | | `/agents spawn <n>` | Spawn parallel agents (1-10) | | `/agents list` | List active agents | | `/agents stop` | Stop agents | | `/kb sync` | Sync to Knowledge Base | | `/tools list` | List available tools | | `/usage` | Token usage | | `/quit` | Exit | ## Architecture ``` +----------------+ | AiCippy CLI | +--------+-------+ | +--------v--------+ | Agent Orchestrator | +--------+---------+ | +--------------------+--------------------+ | | | | | +----v----+ +--v---+ +----v---+ +----v----+ +--v----+ |Agent-1 | |Agent-2| |Agent-3 | |Agent-4 | |Agent-N| |INFRA | |BEDROCK| |API-GW | |CLI-CORE | |... | +---------+ +-------+ +--------+ +---------+ +-------+ | | | | | +----v---------v----------v----------v---------v----+ | AWS Bedrock Runtime | +--------------------------------------------------+ ``` ## Supported Models - **Claude Opus 4.5** (default) - Most capable model - **Claude Sonnet 4.5** - Balanced performance - **Llama 4 Maverick** - Open source alternative ## MCP Tool Connectors - AWS CLI (`aws`) - Google Cloud CLI (`gcloud`) - GitHub CLI (`gh`) - Firebase CLI (`firebase`) - Figma API - Google Drive API - Gmail API - Razorpay API - PayPal API - Stripe CLI - Shell commands (sandboxed) ## Configuration Environment variables (or `.env` file): ```bash AICIPPY_AWS_REGION=us-east-1 AICIPPY_DEFAULT_MODEL=opus AICIPPY_MAX_PARALLEL_AGENTS=10 AICIPPY_LOG_LEVEL=INFO ``` Configuration file: `~/.aicippy/config.toml` ## Security - OAuth 2.0 authentication via AWS Cognito - Tokens stored in OS keychain (macOS Keychain, Windows Credential Manager) - All communications over TLS 1.3 - Secrets never logged or printed - IAM least privilege roles ## Development ```bash # Clone repository git clone https://github.com/aivibe/aicippy.git cd aicippy # Install dev dependencies pip install -e ".[dev]" # Run tests pytest # Run linter ruff check src/ # Type checking mypy src/aicippy/ ``` ## Infrastructure Deployment ```bash # Install CDK dependencies cd infrastructure pip install -r requirements.txt # Deploy all stacks cdk deploy --all ``` ## License Proprietary - AiVibe Software Services Pvt Ltd ## Support - Documentation: https://docs.aicippy.com - Issues: https://github.com/aivibe/aicippy/issues - Email: support@aivibe.in --- Built with precision by AiVibe Software Services Pvt Ltd, Chennai, India.
text/markdown
null
Aravind Jayamohan <aravind@aivibe.in>
null
AiVibe Software Services Pvt Ltd <support@aivibe.in>
Proprietary
ai, aws, bedrock, cli, enterprise, multi-agent
[ "Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: Other/Proprietary License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: System :: Shells", "Typing :: Typed" ]
[]
null
null
>=3.11
[]
[]
[]
[ "aiofiles==25.1.0", "anyio==4.12.1", "asyncpg==0.31.0", "beautifulsoup4==4.14.3", "boto3==1.42.54", "botocore==1.42.54", "feedparser==6.0.12", "httpx==0.28.1", "keyring==25.7.0", "lxml==6.0.2", "markdown==3.10.2", "orjson==3.11.7", "prompt-toolkit==3.0.52", "pydantic-settings==2.13.1", "pydantic==2.12.5", "python-jose[cryptography]==3.5.0", "rich==14.3.3", "structlog==25.5.0", "tenacity==9.1.4", "typer==0.24.0", "websockets==16.0", "aws-cdk-lib==2.239.0; extra == \"cdk\"", "constructs==10.5.1; extra == \"cdk\"", "black==26.1.0; extra == \"dev\"", "boto3-stubs[essential]==1.42.54; extra == \"dev\"", "moto[all]==5.1.21; extra == \"dev\"", "mypy==1.19.1; extra == \"dev\"", "pre-commit==4.5.1; extra == \"dev\"", "pytest-asyncio==1.3.0; extra == \"dev\"", "pytest-cov==7.0.0; extra == \"dev\"", "pytest-mock==3.15.1; extra == \"dev\"", "pytest==9.0.2; extra == \"dev\"", "ruff==0.15.2; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://aicippy.com", "Documentation, https://docs.aicippy.com", "Repository, https://github.com/aivibe/aicippy", "Issues, https://github.com/aivibe/aicippy/issues" ]
twine/6.2.0 CPython/3.14.0
2026-02-20T21:32:50.349749
aicippy-1.6.12.tar.gz
240,803
dd/1a/00b3296fa2c19ea6c2590dc6f35d2325d7443e85bca54961b8d29ae8d28b/aicippy-1.6.12.tar.gz
source
sdist
null
false
a5721ba8ed4f7d70a1b79f6657f89d9e
1c3d252a95f0b32751115b3ae77e42262239954b999a353988d7964d8e58bf2b
dd1a00b3296fa2c19ea6c2590dc6f35d2325d7443e85bca54961b8d29ae8d28b
null
[ "LICENSE" ]
210
2.4
x402
2.2.0
x402 Payment Protocol SDK for Python
# x402 Python SDK Core implementation of the x402 payment protocol. Provides transport-agnostic client, server, and facilitator components with both async and sync variants. ## Installation Install the core package with your preferred framework/client: ```bash # HTTP clients (pick one) uv add x402[httpx] # httpx client uv add x402[requests] # requests client # Server frameworks (pick one) uv add x402[fastapi] # FastAPI middleware uv add x402[flask] # Flask middleware # Blockchain mechanisms (pick one or both) uv add x402[evm] # EVM/Ethereum uv add x402[svm] # Solana # Multiple extras uv add x402[fastapi,httpx,evm] # Everything uv add x402[all] ``` ## Quick Start ### Client (Async) ```python from x402 import x402Client from x402.mechanisms.evm.exact import ExactEvmScheme client = x402Client() client.register("eip155:*", ExactEvmScheme(signer=my_signer)) # Create payment from 402 response payload = await client.create_payment_payload(payment_required) ``` ### Client (Sync) ```python from x402 import x402ClientSync from x402.mechanisms.evm.exact import ExactEvmScheme client = x402ClientSync() client.register("eip155:*", ExactEvmScheme(signer=my_signer)) payload = client.create_payment_payload(payment_required) ``` ### Server (Async) ```python from x402 import x402ResourceServer, ResourceConfig from x402.http import HTTPFacilitatorClient from x402.mechanisms.evm.exact import ExactEvmServerScheme facilitator = HTTPFacilitatorClient(url="https://x402.org/facilitator") server = x402ResourceServer(facilitator) server.register("eip155:*", ExactEvmServerScheme()) server.initialize() # Build requirements config = ResourceConfig( scheme="exact", network="eip155:8453", pay_to="0x...", price="$0.01", ) requirements = server.build_payment_requirements(config) # Verify payment result = await server.verify_payment(payload, requirements[0]) ``` ### Server (Sync) ```python from x402 import x402ResourceServerSync, ResourceConfig from x402.http import HTTPFacilitatorClientSync from x402.mechanisms.evm.exact import ExactEvmServerScheme facilitator = HTTPFacilitatorClientSync(url="https://x402.org/facilitator") server = x402ResourceServerSync(facilitator) server.register("eip155:*", ExactEvmServerScheme()) server.initialize() result = server.verify_payment(payload, requirements[0]) ``` ### Facilitator (Async) ```python from x402 import x402Facilitator from x402.mechanisms.evm.exact import ExactEvmFacilitatorScheme facilitator = x402Facilitator() facilitator.register( ["eip155:8453", "eip155:84532"], ExactEvmFacilitatorScheme(wallet=wallet), ) result = await facilitator.verify(payload, requirements) if result.is_valid: settle_result = await facilitator.settle(payload, requirements) ``` ### Facilitator (Sync) ```python from x402 import x402FacilitatorSync from x402.mechanisms.evm.exact import ExactEvmFacilitatorScheme facilitator = x402FacilitatorSync() facilitator.register( ["eip155:8453", "eip155:84532"], ExactEvmFacilitatorScheme(wallet=wallet), ) result = facilitator.verify(payload, requirements) ``` ## Async vs Sync Each component has both async and sync variants: | Async (default) | Sync | |-----------------|------| | `x402Client` | `x402ClientSync` | | `x402ResourceServer` | `x402ResourceServerSync` | | `x402Facilitator` | `x402FacilitatorSync` | | `HTTPFacilitatorClient` | `HTTPFacilitatorClientSync` | Async variants support both sync and async hooks (auto-detected). Sync variants only support sync hooks and raise `TypeError` if async hooks are registered. ### Framework Pairing | Framework | HTTP Client | Server | Facilitator Client | |-----------|-------------|--------|-------------------| | FastAPI | httpx | `x402ResourceServer` | `HTTPFacilitatorClient` | | Flask | requests | `x402ResourceServerSync` | `HTTPFacilitatorClientSync` | Mismatched variants raise `TypeError` at runtime. ## Client Configuration Use `from_config()` for declarative setup: ```python from x402 import x402Client, x402ClientConfig, SchemeRegistration config = x402ClientConfig( schemes=[ SchemeRegistration(network="eip155:*", client=ExactEvmScheme(signer)), SchemeRegistration(network="solana:*", client=ExactSvmScheme(signer)), ], policies=[prefer_network("eip155:8453")], ) client = x402Client.from_config(config) ``` ## Policies Filter or prioritize payment requirements: ```python from x402 import prefer_network, prefer_scheme, max_amount client.register_policy(prefer_network("eip155:8453")) client.register_policy(prefer_scheme("exact")) client.register_policy(max_amount(1_000_000)) # 1 USDC max ``` ## Lifecycle Hooks ### Client Hooks ```python from x402 import AbortResult, RecoveredPayloadResult def before_payment(ctx): print(f"Creating payment for: {ctx.selected_requirements.network}") # Return AbortResult(reason="...") to cancel def after_payment(ctx): print(f"Payment created: {ctx.payment_payload}") def on_failure(ctx): print(f"Payment failed: {ctx.error}") # Return RecoveredPayloadResult(payload=...) to recover client.on_before_payment_creation(before_payment) client.on_after_payment_creation(after_payment) client.on_payment_creation_failure(on_failure) ``` ### Server Hooks ```python server.on_before_verify(lambda ctx: print(f"Verifying: {ctx.payload}")) server.on_after_verify(lambda ctx: print(f"Result: {ctx.result.is_valid}")) server.on_verify_failure(lambda ctx: print(f"Failed: {ctx.error}")) server.on_before_settle(lambda ctx: ...) server.on_after_settle(lambda ctx: ...) server.on_settle_failure(lambda ctx: ...) ``` ### Facilitator Hooks ```python facilitator.on_before_verify(...) facilitator.on_after_verify(...) facilitator.on_verify_failure(...) facilitator.on_before_settle(...) facilitator.on_after_settle(...) facilitator.on_settle_failure(...) ``` ## Network Pattern Matching Register handlers for network families using wildcards: ```python # All EVM networks client.register("eip155:*", ExactEvmScheme(signer)) # Specific network (takes precedence) client.register("eip155:8453", CustomScheme()) ``` ## HTTP Headers ### V2 Protocol (Current) | Header | Description | |--------|-------------| | `PAYMENT-SIGNATURE` | Base64-encoded payment payload | | `PAYMENT-REQUIRED` | Base64-encoded payment requirements | | `PAYMENT-RESPONSE` | Base64-encoded settlement response | ### V1 Protocol (Legacy) | Header | Description | |--------|-------------| | `X-PAYMENT` | Base64-encoded payment payload | | `X-PAYMENT-RESPONSE` | Base64-encoded settlement response | ## Related Modules - `x402.http` - HTTP clients, middleware, and facilitator client - `x402.mechanisms.evm` - EVM/Ethereum implementation - `x402.mechanisms.svm` - Solana implementation - `x402.extensions` - Protocol extensions (Bazaar discovery) ## Examples See [examples/python](https://github.com/coinbase/x402/tree/main/examples/python).
text/markdown
Coinbase
null
null
null
MIT
402, http, payment, protocol, x402
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Typing :: Typed" ]
[]
null
null
>=3.10
[]
[]
[]
[ "nest-asyncio>=1.6.0", "pydantic>=2.0.0", "typing-extensions>=4.0.0", "eth-abi>=5.0.0; extra == \"all\"", "eth-account>=0.12.0; extra == \"all\"", "eth-keys>=0.5.0; extra == \"all\"", "eth-utils>=4.0.0; extra == \"all\"", "fastapi[standard]>=0.115.0; extra == \"all\"", "flask>=3.0.0; extra == \"all\"", "httpx>=0.28.1; extra == \"all\"", "jsonschema>=4.0.0; extra == \"all\"", "mcp>=1.0.0; extra == \"all\"", "requests>=2.31.0; extra == \"all\"", "solana>=0.36.0; extra == \"all\"", "solders>=0.27.0; extra == \"all\"", "starlette>=0.27.0; extra == \"all\"", "web3>=7.0.0; extra == \"all\"", "httpx>=0.28.1; extra == \"clients\"", "requests>=2.31.0; extra == \"clients\"", "eth-abi>=5.0.0; extra == \"evm\"", "eth-account>=0.12.0; extra == \"evm\"", "eth-keys>=0.5.0; extra == \"evm\"", "eth-utils>=4.0.0; extra == \"evm\"", "web3>=7.0.0; extra == \"evm\"", "jsonschema>=4.0.0; extra == \"extensions\"", "fastapi[standard]>=0.115.0; extra == \"fastapi\"", "starlette>=0.27.0; extra == \"fastapi\"", "flask>=3.0.0; extra == \"flask\"", "httpx>=0.28.1; extra == \"httpx\"", "mcp>=1.0.0; extra == \"mcp\"", "eth-abi>=5.0.0; extra == \"mechanisms\"", "eth-account>=0.12.0; extra == \"mechanisms\"", "eth-keys>=0.5.0; extra == \"mechanisms\"", "eth-utils>=4.0.0; extra == \"mechanisms\"", "solana>=0.36.0; extra == \"mechanisms\"", "solders>=0.27.0; extra == \"mechanisms\"", "web3>=7.0.0; extra == \"mechanisms\"", "requests>=2.31.0; extra == \"requests\"", "fastapi[standard]>=0.115.0; extra == \"servers\"", "flask>=3.0.0; extra == \"servers\"", "starlette>=0.27.0; extra == \"servers\"", "solana>=0.36.0; extra == \"svm\"", "solders>=0.27.0; extra == \"svm\"" ]
[]
[]
[]
[ "Homepage, https://github.com/coinbase/x402", "Documentation, https://x402.org", "Repository, https://github.com/coinbase/x402" ]
twine/6.1.0 CPython/3.12.9
2026-02-20T21:32:28.400161
x402-2.2.0.tar.gz
892,468
97/47/ce13f2f3ab9eed465a4a81009b630df5e7837c34d9c1584c4e5b943b7eb5/x402-2.2.0.tar.gz
source
sdist
null
false
09c68476c6c5f6df42bd68bbb3eac2d3
35e14c80b29b845a7e7d8c394a57dd2ebe2f5bdc9e5e80da219cfd0af688e80f
9747ce13f2f3ab9eed465a4a81009b630df5e7837c34d9c1584c4e5b943b7eb5
null
[]
659
2.4
ninjastack
0.1.0
Ninja Stack — schema-first agentic backend framework
> **⚠️ Early Alpha — Under Heavy Development** > > NinjaStack is actively undergoing heavy development and is **not yet ready for production use**. APIs, schemas, and architecture may change without notice. We're still in early alpha and welcome new contributors and anyone willing to explore a new frontier of agentic backend architecture. If that sounds like your kind of thing — jump in, open issues, and help shape what this becomes. <p align="center"> <h1 align="center">🥷 NinjaStack</h1> <p align="center"> <strong>Schema-first agentic backend framework.</strong><br> Point at a database, get a full agentic backend with AI agents, GraphQL, auth, and UI. </p> <p align="center"> <a href="https://codeninja.github.io/ninja-stack/">Homepage</a> · <a href="https://codeninja.github.io/ninja-stack/docs/">Documentation</a> · <a href="https://codeninja.github.io/ninja-stack/docs/examples/">Examples</a> </p> </p> --- ## What is NinjaStack? NinjaStack transforms database schemas into fully functional agentic backends. Define your data model once — through database introspection or conversational design — and the framework generates AI agents, GraphQL APIs, authentication, RBAC, and deployment manifests. ```bash # Connect to your database, discover the schema ninjastack introspect --db postgres://localhost/myapp # Generate everything: models, agents, GraphQL, auth ninjastack sync # Run your agentic backend ninjastack serve # → Agentic backend at http://localhost:8000 # → GraphQL playground at /graphql # → Agent chat at /chat ``` No database yet? Chat with the AI setup assistant to design your schema through natural dialogue: ```bash ninjastack init --interactive # "I need a bookstore with books, customers, orders, and reviews..." ``` ## Key Features | Feature | Description | |---------|-------------| | 🔍 **Database Introspection** | Auto-discover entities from PostgreSQL, MongoDB, Neo4j, or vector stores | | 🤖 **ADK Agent Generation** | Google ADK agents with scoped CRUD tools per entity | | 🧬 **Agentic Schema Definition** | Typed, composable schema language — your single source of truth | | 🔐 **Auth & RBAC** | Pluggable auth (OAuth2, JWT, API keys) with declarative role-based permissions | | 📊 **GraphQL Generation** | Strawberry types, queries, and mutations from schema | | 💬 **Conversational Setup** | Design your schema through natural dialogue with Gemini | | 🎯 **Tool Scoping** | Each agent only sees its own tools — no leaking across boundaries | | 🚀 **K8s Deployment** | Helm charts and manifests generated automatically | | 🔄 **Polyglot Persistence** | Unified layer across SQL, NoSQL, graph, and vector databases | ## Architecture NinjaStack organizes agents in a three-tier hierarchy with explicit ownership at every level: ```mermaid graph TD C["🎯 Coordinator Agent<br/><small>LLM · gemini-2.5-pro · Intent routing</small>"] C --> D1["📚 Catalog Domain<br/><small>gemini-2.5-flash · Medium reasoning</small>"] C --> D2["🛒 Commerce Domain<br/><small>gemini-2.5-pro · High reasoning</small>"] D1 --> B["📖 Book Agent"] D1 --> R["⭐ Review Agent"] D2 --> Cu["👤 Customer Agent"] D2 --> O["📦 Order Agent"] B --> P["🗄️ Unified Persistence Layer<br/><small>SQL · MongoDB · Neo4j · ChromaDB</small>"] R --> P Cu --> P O --> P style C fill:#166534,color:#fff,stroke:#22c55e style D1 fill:#1e3a5f,color:#fff,stroke:#3b82f6 style D2 fill:#1e3a5f,color:#fff,stroke:#3b82f6 style B fill:#854d0e,color:#fff,stroke:#eab308 style R fill:#854d0e,color:#fff,stroke:#eab308 style Cu fill:#854d0e,color:#fff,stroke:#eab308 style O fill:#854d0e,color:#fff,stroke:#eab308 style P fill:#581c87,color:#fff,stroke:#a855f7 ``` - **Data Agents** — Deterministic CRUD. No LLM. One entity, scoped tools. Fast and testable. - **Domain Agents** — LLM-powered. Own a business domain. Delegate to data agents. Configurable reasoning. - **Coordinator** — Top-level router. Classifies intent. Synthesizes cross-domain results. > 📚 [Full architecture docs →](https://codeninja.github.io/ninja-stack/docs/architecture/) ## Quick Start ### Prerequisites - Python 3.12+ - [uv](https://docs.astral.sh/uv/) package manager ### Install from source ```bash git clone https://github.com/codeninja/ninja-stack.git cd ninja-stack uv sync ``` ### Run the examples All examples use a bookstore domain and work without any API keys: ```bash # Schema definition PYTHONPATH=examples/bookstore uv run python examples/bookstore/01_schema_definition.py # Data agents (deterministic CRUD) PYTHONPATH=examples/bookstore uv run python examples/bookstore/03_data_agents.py # Full end-to-end pipeline PYTHONPATH=examples/bookstore uv run python examples/bookstore/06_end_to_end.py ``` | # | Example | What It Demonstrates | |---|---------|---------------------| | 1 | [Schema Definition](examples/bookstore/01_schema_definition.py) | Entities, fields, relationships, domains | | 2 | [Code Generation](examples/bookstore/02_code_generation.py) | Generate models, agents, GraphQL from schema | | 3 | [Data Agents](examples/bookstore/03_data_agents.py) | Deterministic CRUD, tool scoping, tracing | | 4 | [Domain Agents](examples/bookstore/04_domain_agents.py) | LLM-powered orchestration and delegation | | 5 | [Auth & RBAC](examples/bookstore/05_auth_rbac.py) | Identity, JWT tokens, role-based permissions | | 6 | [End-to-End](examples/bookstore/06_end_to_end.py) | Full pipeline: schema → agents → auth → query | ### Optional: Enable LLM features Data agents, code generation, and RBAC work without an API key. For LLM-powered features (domain agents, conversational setup): ```bash export GOOGLE_API_KEY="your-gemini-api-key" ``` ## Project Structure NinjaStack is a modular monorepo of 15 focused packages: ``` ninja-stack/ ├── libs/ # Reusable libraries │ ├── ninja-core/ # ASD schema models (entity, domain, relationship) │ ├── ninja-agents/ # ADK agents (DataAgent, DomainAgent, Coordinator) │ ├── ninja-auth/ # Auth gateway, strategies, RBAC │ ├── ninja-codegen/ # Jinja2 code generation engine │ ├── ninja-introspect/ # Database schema discovery │ ├── ninja-persistence/ # Unified polyglot persistence │ ├── ninja-gql/ # Strawberry GraphQL generation │ ├── ninja-boundary/ # Data tolerance & coercion │ ├── ninja-graph/ # Graph-RAG bootstrapper │ ├── ninja-models/ # Pydantic model generation │ ├── ninja-deploy/ # K8s/Helm deployment pipeline │ ├── ninja-ui/ # CRUD viewer & chat UI generation │ └── ninja-cli/ # CLI tooling ├── apps/ # Deployable applications │ ├── ninja-api/ # FastAPI server │ └── ninja-setup-assistant/ # Gemini-powered conversational setup ├── examples/ # Bookstore walkthrough (6 examples) ├── docs/ # MkDocs source └── site/ # Landing page + built docs ``` ## Tech Stack | Layer | Technology | |-------|-----------| | Language | Python 3.12+ · Pydantic v2 | | Agents | Google ADK · LiteLLM (model-agnostic) | | API | FastAPI · Strawberry GraphQL | | Auth | JWT · OAuth2 · API Keys · bcrypt | | Persistence | SQLAlchemy · Motor/Beanie · Neo4j · ChromaDB | | Deploy | Kubernetes · Helm | | Package Mgmt | uv | ## Contributing ### Setup ```bash git clone https://github.com/codeninja/ninja-stack.git cd ninja-stack uv sync ``` ### Run tests ```bash # Full suite uv run pytest # Specific library uv run pytest libs/ninja-core/ uv run pytest libs/ninja-agents/ uv run pytest libs/ninja-auth/ # With coverage uv run pytest --cov ``` ### Project conventions - **Commits**: [Conventional Commits](https://www.conventionalcommits.org/) — `feat(ninja-agents): add tool scoping` - **Branches**: `feat/issue-<N>-description` from `main` - **PRs**: One feature per PR, linked to an issue - **Tests**: Every library has its own test suite. All tests must pass before merge. - **Code generation**: Templates live in `libs/ninja-codegen/src/ninja_codegen/templates/` - **Adding a library**: Create under `libs/`, add to root `pyproject.toml` workspace members ### Build docs locally ```bash uv run mkdocs serve # → http://localhost:8000 ``` ## Links - 🏠 **Homepage**: [codeninja.github.io/ninja-stack](https://codeninja.github.io/ninja-stack/) - 📚 **Documentation**: [codeninja.github.io/ninja-stack/docs](https://codeninja.github.io/ninja-stack/docs/) - 📖 **Examples**: [examples/bookstore/](examples/bookstore/) - 🐛 **Issues**: [github.com/codeninja/ninja-stack/issues](https://github.com/codeninja/ninja-stack/issues) ## License TBD
text/markdown
null
null
null
null
null
null
[]
[]
null
null
>=3.12
[]
[]
[]
[ "aiosqlite>=0.20", "bcrypt>=4.0", "google-adk>=0.4.0", "httpx>=0.25", "jinja2>=3.1", "litellm>=1.0", "pydantic>=2.0", "pyjwt[crypto]>=2.8", "pyyaml>=6.0", "sqlalchemy[asyncio]>=2.0", "starlette>=0.27", "strawberry-graphql>=0.236", "typer>=0.15", "chromadb>=0.4; extra == \"all\"", "motor>=3.0; extra == \"all\"", "neo4j>=5.0; extra == \"all\"", "pymilvus>=2.3; extra == \"all\"", "neo4j>=5.0; extra == \"graph\"", "motor>=3.0; extra == \"mongo\"", "chromadb>=0.4; extra == \"vector\"", "pymilvus>=2.3; extra == \"vector\"" ]
[]
[]
[]
[]
uv/0.8.23
2026-02-20T21:32:09.848396
ninjastack-0.1.0.tar.gz
1,371,142
36/0b/2ca046908e60b12c6efd69aea3140f648ccc7cde970a299ae7def657871b/ninjastack-0.1.0.tar.gz
source
sdist
null
false
9d87e4f0559cedc2d0b2badaac968554
c370992740ecef6fa281b5797a1d34c7eed5380b1ad87f0ddfd76e57be68bac6
360b2ca046908e60b12c6efd69aea3140f648ccc7cde970a299ae7def657871b
null
[]
214