metadata_version
string
name
string
version
string
summary
string
description
string
description_content_type
string
author
string
author_email
string
maintainer
string
maintainer_email
string
license
string
keywords
string
classifiers
list
platform
list
home_page
string
download_url
string
requires_python
string
requires
list
provides
list
obsoletes
list
requires_dist
list
provides_dist
list
obsoletes_dist
list
requires_external
list
project_urls
list
uploaded_via
string
upload_time
timestamp[us]
filename
string
size
int64
path
string
python_version
string
packagetype
string
comment_text
string
has_signature
bool
md5_digest
string
sha256_digest
string
blake2_256_digest
string
license_expression
string
license_files
list
recent_7d_downloads
int64
2.3
casbin-fastapi-decorator-jwt
0.1.4
JWT user provider for casbin-fastapi-decorator
# casbin-fastapi-decorator-jwt JWT user provider for [casbin-fastapi-decorator](https://github.com/Neko1313/casbin-fastapi-decorator). Extracts and validates a JWT from the Bearer header and/or a cookie, returning the payload as the current user. ## Installation ```bash pip install casbin-fastapi-decorator-jwt ``` Or via the core package extra: ```bash pip install "casbin-fastapi-decorator[jwt]" ``` ## Usage ```python from casbin_fastapi_decorator_jwt import JWTUserProvider user_provider = JWTUserProvider( secret_key="your-secret", algorithm="HS256", # default cookie_name="access_token", # optional, enables reading from cookie user_model=UserSchema, # optional, Pydantic model for payload validation ) ``` Pass it to `PermissionGuard` as the `user_provider`: ```python import casbin from fastapi import FastAPI, HTTPException from casbin_fastapi_decorator import PermissionGuard async def get_enforcer() -> casbin.Enforcer: return casbin.Enforcer("model.conf", "policy.csv") guard = PermissionGuard( user_provider=user_provider, enforcer_provider=get_enforcer, error_factory=lambda user, *rv: HTTPException(403, "Forbidden"), ) app = FastAPI() @app.get("/articles") @guard.require_permission("articles", "read") async def list_articles(): return [] ``` ## API ### `JWTUserProvider` ```python JWTUserProvider( secret_key: str, algorithm: str = "HS256", cookie_name: str | None = None, user_model: type[BaseModel] | None = None, ) ``` | Parameter | Description | |---|---| | `secret_key` | Secret used to verify the JWT signature | | `algorithm` | JWT algorithm (default: `"HS256"`) | | `cookie_name` | If set, also reads the token from this cookie name | | `user_model` | Pydantic model — if provided, the payload is validated via `model_validate()` | When called as a FastAPI dependency, the provider reads the token from: 1. `Authorization: Bearer <token>` header (always) 2. Cookie `<cookie_name>` (if `cookie_name` is set) ## Development See the [workspace README](../../README.md) for setup instructions. ```bash task jwt:lint # ruff + bandit + ty task jwt:test # pytest ```
text/markdown
null
null
null
null
null
null
[]
[]
null
null
>=3.10
[]
[]
[]
[ "casbin-fastapi-decorator", "pyjwt>=2.0.0" ]
[]
[]
[]
[]
uv/0.10.4 {"installer":{"name":"uv","version":"0.10.4","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
2026-02-21T01:11:49.752469
casbin_fastapi_decorator_jwt-0.1.4-py3-none-any.whl
4,103
64/8d/fcee5effa2311491d33b7517dc2ee995a6490dd9c427e977d44eef296c60/casbin_fastapi_decorator_jwt-0.1.4-py3-none-any.whl
py3
bdist_wheel
null
false
9cf0e7dd9491d21a22af117f748aa278
9423bfb36051eaa93edc1ded3896734aecf16b2db1abdb7fb1a93d11aa175d78
648dfcee5effa2311491d33b7517dc2ee995a6490dd9c427e977d44eef296c60
null
[]
229
2.3
casbin-fastapi-decorator
0.1.4
Casbin authorization decorator factory for FastAPI
# casbin-fastapi-decorator Authorization decorator factory for FastAPI based on [Casbin](https://casbin.org/) and [fastapi-decorators](https://pypi.org/project/fastapi-decorators/). Decorators are applied to routes — no middleware or dependencies in the endpoint signature. ## Installation ```bash pip install casbin-fastapi-decorator ``` Additional providers: ```bash pip install "casbin-fastapi-decorator[jwt]" # JWT authentication pip install "casbin-fastapi-decorator[db]" # Policies from DB (SQLAlchemy) ``` ## Quick start ```python import casbin from fastapi import FastAPI, HTTPException from casbin_fastapi_decorator import AccessSubject, PermissionGuard # 1. Providers — regular FastAPI dependencies async def get_current_user() -> dict: return {"sub": "alice", "role": "admin"} async def get_enforcer() -> casbin.Enforcer: return casbin.Enforcer("model.conf", "policy.csv") # 2. Decorator factory guard = PermissionGuard( user_provider=get_current_user, enforcer_provider=get_enforcer, error_factory=lambda user, *rv: HTTPException(403, "Forbidden"), ) app = FastAPI() # 3. Authentication only @app.get("/me") @guard.auth_required() async def me(): return {"ok": True} # 4. Static permission check @app.get("/articles") @guard.require_permission("articles", "read") async def list_articles(): return [] # 5. Dynamic check — value from request async def get_article(article_id: int) -> dict: return {"id": article_id, "owner": "alice"} @app.get("/articles/{article_id}") @guard.require_permission( AccessSubject(val=get_article, selector=lambda a: a["owner"]), "read", ) async def read_article(article_id: int): return {"article_id": article_id} ``` Arguments of `require_permission` are passed to `enforcer.enforce(user, *args)` in the same order. `AccessSubject` is resolved via FastAPI DI, then transformed by the `selector`. ## API ### `PermissionGuard` ```python PermissionGuard( user_provider=..., # FastAPI dependency that returns the current user enforcer_provider=..., # FastAPI dependency that returns a casbin.Enforcer error_factory=..., # callable(user, *rvals) -> Exception ) ``` | Method | Description | |---|---| | `auth_required()` | Decorator: authentication only (user_provider must not raise an exception) | | `require_permission(*args)` | Decorator: permission check via `enforcer.enforce(user, *args)` | ### `AccessSubject` ```python AccessSubject( val=get_item, # FastAPI dependency selector=lambda item: item["name"], # transformation before enforce ) ``` Wraps a dependency whose value needs to be obtained from the request and passed to the enforcer. By default, `selector` is identity (`lambda x: x`). ## JWT provider [`casbin-fastapi-decorator-jwt`](packages/casbin-fastapi-decorator-jwt) — extracts and validates a JWT from the Bearer header and/or a cookie. ```bash pip install "casbin-fastapi-decorator[jwt]" ``` See [packages/casbin-fastapi-decorator-jwt/README.md](packages/casbin-fastapi-decorator-jwt/README.md) for full API and usage. ## DB provider [`casbin-fastapi-decorator-db`](packages/casbin-fastapi-decorator-db) — loads Casbin policies from a SQLAlchemy async session. ```bash pip install "casbin-fastapi-decorator[db]" ``` See [packages/casbin-fastapi-decorator-db/README.md](packages/casbin-fastapi-decorator-db/README.md) for full API and usage. ## Examples | Example | Description | |---|---| | [`examples/core`](examples/core) | Bearer token auth, file-based Casbin policies | | [`examples/core-jwt`](examples/core-jwt) | JWT auth via `JWTUserProvider`, file-based policies | | [`examples/core-db`](examples/core-db) | Bearer token auth, policies from SQLite via `DatabaseEnforcerProvider` | ## Development Requires Python 3.10+, [uv](https://docs.astral.sh/uv/), [task](https://taskfile.dev/). ```bash task install # uv sync --all-groups + install extras (jwt, db) task lint # ruff + ty + bandit for all packages task tests # all tests (core + jwt + db) ``` Individual package tasks: ```bash task core:lint # lint core only task core:test # test core only task jwt:lint # lint JWT package task jwt:test # test JWT package task db:lint # lint DB package task db:test # test DB package (requires Docker for testcontainers) ``` ## License MIT
text/markdown
Neko1313
Neko1313 <nikita.ribalchencko@yandex.ru>
null
null
MIT License Copyright (c) 2026 Neko1313 Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
casbin, fastapi, authorization, rbac, abac, permissions, decorator, access-control, jwt, sqlalchemy
[ "Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3 :: Only", "Framework :: FastAPI", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: HTTP Servers", "Topic :: Security", "Topic :: Software Development :: Libraries :: Python Modules", "Typing :: Typed", "Operating System :: OS Independent" ]
[]
null
null
>=3.10
[]
[]
[]
[ "fastapi>=0.115.0", "fastapi-decorators>=0.5.0", "casbin>=1.36.0", "casbin-fastapi-decorator-db; extra == \"db\"", "casbin-fastapi-decorator-jwt; extra == \"jwt\"" ]
[]
[]
[]
[]
uv/0.10.4 {"installer":{"name":"uv","version":"0.10.4","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
2026-02-21T01:11:48.851927
casbin_fastapi_decorator-0.1.4-py3-none-any.whl
6,574
70/fd/6eca16ac606257173e2a2a0ea7e431561ac984fc38ba070ea24a66f0467f/casbin_fastapi_decorator-0.1.4-py3-none-any.whl
py3
bdist_wheel
null
false
987281b4c14483f68ee56468a876cd1f
2da506d4bb17f83dd642cfc5b922c723e0a7c84b7c3e04d9273ddfa94fc622b5
70fd6eca16ac606257173e2a2a0ea7e431561ac984fc38ba070ea24a66f0467f
null
[]
260
2.3
gen-dsp
0.1.12
Generate multiple dsp plugin formats from Max gen~ exports
# gen-dsp gen-dsp is a zero-dependency pure Python package that generates buildable audio plugin projects from Max/MSP gen~ code exports, targeting PureData, Max/MSP, ChucK, AudioUnit (AUv2), CLAP, VST3, LV2, SuperCollider, VCV Rack, Daisy, and Circle. It handles project scaffolding, I/O and buffer detection, parameter metadata extraction, and platform-specific patching. This project is a friendly fork of Michael Spears' [gen_ext](https://github.com/samesimilar/gen_ext) which was originally created to "compile code exported from a Max gen~ object into an "external" object that can be loaded into a PureData patch." ## Cross-Platform Support gen-dsp builds on macOS, Linux, and Windows. All platforms are tested in CI via GitHub Actions. | Platform | macOS | Linux | Windows | Build System | Output | |----------|:-----:|:-----:|:-------:|--------------|--------| | PureData | yes | yes | -- | make (pd-lib-builder) | `.pd_darwin` / `.pd_linux` | | Max/MSP | yes | -- | -- | CMake (max-sdk-base) | `.mxo` / `.mxe64` | | ChucK | yes | yes | -- | make | `.chug` | | AudioUnit | yes | -- | -- | CMake | `.component` | | CLAP | yes | yes | yes | CMake (FetchContent) | `.clap` | | VST3 | yes | yes | yes | CMake (FetchContent) | `.vst3` | | LV2 | yes | yes | -- | CMake (FetchContent) | `.lv2` | | SuperCollider | yes | yes | yes | CMake (FetchContent) | `.scx` / `.so` | | VCV Rack | yes | yes | -- | make (Rack SDK) | `plugin.dylib` / `.so` / `.dll` | | Daisy | -- | yes | -- | make (libDaisy) | `.bin` (firmware) | | Circle | -- | yes | -- | make (Circle SDK) | `.img` (kernel image) | Each platform has a detailed guide covering prerequisites, build details, SDK configuration, install paths, and troubleshooting: | Platform | Guide | |----------|-------| | PureData | [docs/backends/puredata.md](docs/backends/puredata.md) | | Max/MSP | [docs/backends/max.md](docs/backends/max.md) | | ChucK | [docs/backends/chuck.md](docs/backends/chuck.md) | | AudioUnit (AUv2) | [docs/backends/audiounit.md](docs/backends/audiounit.md) | | CLAP | [docs/backends/clap.md](docs/backends/clap.md) | | VST3 | [docs/backends/vst3.md](docs/backends/vst3.md) | | LV2 | [docs/backends/lv2.md](docs/backends/lv2.md) | | SuperCollider | [docs/backends/supercollider.md](docs/backends/supercollider.md) | | VCV Rack | [docs/backends/vcvrack.md](docs/backends/vcvrack.md) | | Daisy | [docs/backends/daisy.md](docs/backends/daisy.md) | | Circle | [docs/backends/circle.md](docs/backends/circle.md) | ## Key Improvements and Features - **Python package**: gen-dsp is a pip installable zero-dependency python package with a cli which embeds all templates and related code. - **Automated project scaffolding**: `gen-dsp init` creates a complete, buildable project from a gen~ export in one command, versus manually copying files and editing Makefiles. - **Automatic buffer detection**: Scans exported code for buffer usage patterns and configures them without manual intervention. - **Max/MSP support**: Generates CMake-based Max externals with proper 64-bit signal handling and buffer lock/unlock API. - **ChucK support**: Generates chugins (.chug) with multi-channel I/O and runtime parameter control. - **AudioUnit support**: Generates macOS AUv2 plugins (.component) using the raw C API -- no Apple SDK dependency, just system frameworks. - **CLAP support**: Generates cross-platform CLAP plugins (`.clap`) with zero-copy audio processing -- CLAP headers fetched via CMake FetchContent. - **VST3 support**: Generates cross-platform VST3 plugins (`.vst3`) with zero-copy audio processing -- Steinberg VST3 SDK fetched via CMake FetchContent. - **LV2 support**: Generates cross-platform LV2 plugins (`.lv2` bundles) with TTL metadata containing real parameter names/ranges parsed from gen~ exports -- LV2 headers fetched via CMake FetchContent. - **SuperCollider support**: Generates cross-platform SC UGens (`.scx`/`.so`) with `.sc` class files containing parameter names/defaults parsed from gen~ exports -- SC plugin headers fetched via CMake FetchContent. - **VCV Rack support**: Generates VCV Rack modules with per-sample processing, auto-generated `plugin.json` manifest and panel SVG -- Rack SDK auto-downloaded and cached on first build. - **Daisy support**: Generates Daisy Seed firmware (.bin) with custom genlib runtime (bump allocator for SRAM/SDRAM) -- first embedded/cross-compilation target, requires `arm-none-eabi-gcc`. - **Circle support**: Generates bare-metal Raspberry Pi kernel images (.img) for Pi Zero through Pi 5 using the [Circle](https://github.com/rsta2/circle) framework -- 14 board variants covering I2S, PWM, HDMI, and USB audio outputs. Supports multi-plugin mode via `--graph` with USB MIDI CC parameter control: linear chains use an optimized ping-pong buffer codegen path, while arbitrary DAGs (fan-out, fan-in via mixer nodes) use topological sort with per-edge buffer allocation. - **Platform-specific patches**: Automatically fixes compatibility issues like the `exp2f -> exp2` problem in Max 9 exports on macOS. - **Analysis tools**: `gen-dsp detect` inspects exports to show I/O counts, parameters, and buffers before committing to a build. - **Dry-run mode**: Preview what changes will be made before applying them. - **Platform registry**: To make it easy to discover new backends ## Installation ```bash pip install gen-dsp ``` Or install from source: ```bash git clone https://github.com/shakfu/gen-dsp.git cd gen-dsp pip install -e . ``` ## Quick Start ```bash # 1. Export your gen~ code in Max (send 'exportcode' to gen~ object) # 2. Create a project from the export gen-dsp init ./path/to/export -n myeffect -o ./myeffect # 3. Build the external cd myeffect make all # 4. Use in PureData as myeffect~ ``` ## Commands ### init Create a new project from a gen~ export: ```bash gen-dsp init <export-path> -n <name> [-p <platform>] [-o <output>] ``` Options: - `-n, --name` - Name for the external (required) - `-p, --platform` - Target platform: `pd` (default), `max`, `chuck`, `au`, `clap`, `vst3`, `lv2`, `sc`, `vcvrack`, `daisy`, `circle`, or `both` - `-o, --output` - Output directory (default: `./<name>`) - `--buffers` - Explicit buffer names (overrides auto-detection) - `--shared-cache` - Use a shared OS cache for FetchContent downloads (clap, vst3, lv2, sc only) - `--board` - Board variant for embedded platforms (Daisy: `seed`, `pod`, etc.; Circle: `pi3-i2s`, `pi4-usb`, etc.) - `--graph` - JSON graph file for multi-plugin chain mode (Circle only; see [Chain Mode](#circle-chain-mode) below) - `--export` - Additional export path for chain node resolution (repeatable; use with `--graph`) - `--no-patch` - Skip automatic exp2f fix - `--dry-run` - Preview without creating files ### build Build an existing project: ```bash gen-dsp build [project-path] [-p <platform>] [--clean] [-v] ``` ### manifest Emit a JSON manifest describing a gen~ export (I/O counts, parameters with ranges, buffers): ```bash gen-dsp manifest <export-path> [--buffers sample envelope] ``` The same manifest is also written as `manifest.json` to the project root during `gen-dsp init`. ### detect Analyze a gen~ export: ```bash gen-dsp detect <export-path> [--json] ``` Shows: export name, signal I/O counts, parameters, detected buffers, and needed patches. ### patch Apply platform-specific fixes: ```bash gen-dsp patch <target-path> [--dry-run] ``` Currently applies the `exp2f -> exp2` fix for macOS compatibility with Max 9 exports. ## Features ### Automatic Buffer Detection gen-dsp scans your gen~ export for buffer usage patterns and configures them automatically: ```bash $ gen-dsp detect ./my_sampler_export Gen~ Export: my_sampler Signal inputs: 1 Signal outputs: 2 Parameters: 3 Buffers: ['sample', 'envelope'] ``` Buffer names must be valid C identifiers (alphanumeric, starting with a letter). ### Platform Patches Max 9 exports include `exp2f` which fails on macOS. gen-dsp automatically patches this to `exp2` during project creation, or you can apply it manually: ```bash gen-dsp patch ./my_project --dry-run # Preview gen-dsp patch ./my_project # Apply ``` ## PureData See the [PureData guide](docs/backends/puredata.md) for full details. ```bash gen-dsp init ./my_export -n myeffect -p pd -o ./myeffect_pd cd myeffect_pd && make all ``` Parameters: send `<name> <value>` messages to the first inlet. Send `bang` to list all parameters. Buffers connect to PureData arrays by name; use `pdset` to remap. ## Max/MSP See the [Max/MSP guide](docs/backends/max.md) for full details. ```bash gen-dsp init ./my_export -n myeffect -p max -o ./myeffect_max gen-dsp build ./myeffect_max -p max # Output: externals/myeffect~.mxo (macOS) or myeffect~.mxe64 (Windows) ``` Max is the only platform using 64-bit double signals. The SDK (max-sdk-base) is auto-cloned on first build. ## ChucK See the [ChucK guide](docs/backends/chuck.md) for full details. ```bash gen-dsp init ./my_export -n myeffect -p chuck -o ./myeffect_chuck cd myeffect_chuck && make mac # or make linux ``` Class names are auto-capitalized (`myeffect` -> `Myeffect`). Parameters are controlled via `eff.param("name", value)`. Buffer-based chugins can load WAV files at runtime via `eff.loadBuffer("sample", "amen.wav")`. ## AudioUnit (AUv2) See the [AudioUnit guide](docs/backends/audiounit.md) for full details. ```bash gen-dsp init ./my_export -n myeffect -p au -o ./myeffect_au cd myeffect_au && cmake -B build && cmake --build build # Output: build/myeffect.component ``` macOS only. Uses the raw AUv2 C API -- no external SDK needed, just system frameworks. Auto-detects `aufx` (effect) vs `augn` (generator). Passes Apple's `auval` validation. ## CLAP See the [CLAP guide](docs/backends/clap.md) for full details. ```bash gen-dsp init ./my_export -n myeffect -p clap -o ./myeffect_clap cd myeffect_clap && cmake -B build && cmake --build build # Output: build/myeffect.clap ``` Cross-platform (macOS, Linux, Windows). Zero-copy audio. Passes [clap-validator](https://github.com/free-audio/clap-validator) conformance tests. CLAP headers fetched via CMake FetchContent (tag 1.2.2, MIT licensed). ## VST3 See the [VST3 guide](docs/backends/vst3.md) for full details. ```bash gen-dsp init ./my_export -n myeffect -p vst3 -o ./myeffect_vst3 cd myeffect_vst3 && cmake -B build && cmake --build build # Output: build/VST3/Release/myeffect.vst3/ ``` Cross-platform (macOS, Linux, Windows). Zero-copy audio. Passes Steinberg's SDK validator (47/47 tests). VST3 SDK fetched via CMake FetchContent (tag v3.7.9_build_61, GPL3/proprietary dual licensed). ## LV2 See the [LV2 guide](docs/backends/lv2.md) for full details. ```bash gen-dsp init ./my_export -n myeffect -p lv2 -o ./myeffect_lv2 cd myeffect_lv2 && cmake -B build && cmake --build build # Output: build/myeffect.lv2/ ``` macOS and Linux. Passes lilv-based instantiation and audio processing validation. LV2 headers fetched via CMake FetchContent (tag v1.18.10, ISC licensed). TTL metadata with real parameter names/ranges generated at project creation time. ## SuperCollider See the [SuperCollider guide](docs/backends/supercollider.md) for full details. ```bash gen-dsp init ./my_export -n myeffect -p sc -o ./myeffect_sc cd myeffect_sc && cmake -B build && cmake --build build # Output: build/myeffect.scx (macOS) or build/myeffect.so (Linux) ``` Cross-platform (macOS, Linux, Windows). Passes sclang class compilation and scsynth NRT audio rendering validation. SC plugin headers fetched via CMake FetchContent (~80MB tarball). Generates `.sc` class file with parameter names/defaults. UGen name is auto-capitalized. ## VCV Rack See the [VCV Rack guide](docs/backends/vcvrack.md) for full details. ```bash gen-dsp init ./my_export -n myeffect -p vcvrack -o ./myeffect_vcvrack cd myeffect_vcvrack && make # Rack SDK auto-downloaded # Output: plugin.dylib (macOS), plugin.so (Linux), or plugin.dll (Windows) ``` Per-sample processing via `perform(n=1)`. Auto-generates `plugin.json` manifest and dark panel SVG. Passes headless Rack runtime validation (plugin loading + module instantiation). Rack SDK v2.6.1 auto-downloaded and cached. ## Daisy (Electrosmith) See the [Daisy guide](docs/backends/daisy.md) for full details. ```bash gen-dsp init ./my_export -n myeffect -p daisy -o ./myeffect_daisy gen-dsp build ./myeffect_daisy -p daisy # Output: build/myeffect.bin ``` Cross-compilation target for STM32H750. Requires `arm-none-eabi-gcc`. libDaisy (v7.1.0) auto-cloned on first build. Supports 8 board variants via `--board` flag (seed, pod, patch, patch_sm, field, petal, legio, versio). ## Circle (Raspberry Pi bare metal) See the [Circle guide](docs/backends/circle.md) for full details. ```bash gen-dsp init ./my_export -n myeffect -p circle --board pi3-i2s -o ./myeffect_circle gen-dsp build ./myeffect_circle -p circle # Output: kernel8.img (copy to SD card boot partition) ``` Bare-metal kernel images for Raspberry Pi using the [Circle](https://github.com/rsta2/circle) framework (no OS). Requires `aarch64-none-elf-gcc` (64-bit) or `arm-none-eabi-gcc` (32-bit Pi Zero). Circle SDK auto-cloned on first build. Supports 14 board variants via `--board` flag: | Audio | Boards | |-------|--------| | I2S (external DAC) | `pi0-i2s`, `pi0w2-i2s`, `pi3-i2s` (default), `pi4-i2s`, `pi5-i2s` | | PWM (3.5mm jack) | `pi0-pwm`, `pi0w2-pwm`, `pi3-pwm`, `pi4-pwm` | | HDMI | `pi3-hdmi`, `pi4-hdmi`, `pi5-hdmi` | | USB (USB DAC) | `pi4-usb`, `pi5-usb` | ### Circle Multi-Plugin Mode Multi-plugin mode lets you run multiple gen~ plugins on a single Circle kernel image, with USB MIDI CC parameter control at runtime. Provide a JSON graph file via `--graph`: ```bash gen-dsp init ./exports -n mychain -p circle --graph chain.json --board pi4-i2s -o ./mychain cd mychain && make # Output: kernel8-rpi4.img ``` #### Linear chain (auto-detected) When connections form a simple series, gen-dsp uses an optimized ping-pong buffer codegen path: ```json { "nodes": { "reverb": { "export": "gigaverb", "midi_channel": 1 }, "comp": { "export": "compressor", "midi_channel": 2, "cc": { "21": "threshold", "22": "ratio" } } }, "connections": [ ["audio_in", "reverb"], ["reverb", "comp"], ["comp", "audio_out"] ] } ``` #### DAG with fan-out and mixer For non-linear topologies, use fan-out (one output feeding multiple nodes) and built-in mixer nodes for fan-in: ```json { "nodes": { "reverb": { "export": "gigaverb" }, "delay": { "export": "spectraldelayfb" }, "mix": { "type": "mixer", "inputs": 2 } }, "connections": [ ["audio_in", "reverb"], ["audio_in", "delay"], ["reverb", "mix:0"], ["delay", "mix:1"], ["mix", "audio_out"] ] } ``` Mixer nodes combine inputs via weighted sum with per-input gain parameters (default 1.0, range 0.0-2.0), controllable via MIDI CC like any other parameter. The `"mix:0"` syntax routes to a specific mixer input index. #### Graph reference - **`nodes`**: dict of `id -> config`. Gen~ nodes require `"export"` (references a gen~ export directory name under the base `export_path`). Mixer nodes require `"type": "mixer"` and `"inputs"` count. `midi_channel` defaults to position + 1. `cc` is optional (default: CC-by-param-index). - **`connections`**: list of `[from, to]` pairs. `audio_in` and `audio_out` are reserved endpoints. Use `"node:N"` to target a specific input index on mixer nodes. The positional `export_path` argument is the base directory; each node's `export` field is resolved as a subdirectory (e.g. `./exports/gigaverb/gen/`). Use `--export /path/to/export` to provide explicit paths for individual nodes. At runtime, connect a USB MIDI controller. Each node listens on its assigned MIDI channel for CC messages. With the default CC-by-index mapping, CC 0 controls parameter 0, CC 1 controls parameter 1, etc. Explicit `cc` mappings let you assign specific CC numbers to named parameters. ## Shared FetchContent Cache CLAP, VST3, LV2, and SC backends use CMake FetchContent to download their SDKs/headers at configure time. By default each project downloads its own copy. Two opt-in mechanisms allow sharing a single download across projects: ### `--shared-cache` flag Pass `--shared-cache` to `gen-dsp init` to bake an OS-appropriate cache path into the generated CMakeLists.txt: ```bash gen-dsp init ./my_export -n myeffect -p vst3 --shared-cache ``` This resolves to: | OS | Cache path | |----|------------| | macOS | `~/Library/Caches/gen-dsp/fetchcontent/` | | Linux | `$XDG_CACHE_HOME/gen-dsp/fetchcontent/` (defaults to `~/.cache/`) | | Windows | `%LOCALAPPDATA%/gen-dsp/fetchcontent/` | ### `GEN_DSP_CACHE_DIR` environment variable Set this at cmake configure time to override any baked-in path (or use it without `--shared-cache`): ```bash GEN_DSP_CACHE_DIR=/path/to/cache cmake .. ``` The env var takes highest priority, followed by the `--shared-cache` path, followed by CMake's default (project-local `build/_deps/`). The development Makefile exports `GEN_DSP_CACHE_DIR=build/.fetchcontent_cache` automatically, so `make example-clap`, `make example-vst3`, `make example-lv2`, and `make example-sc` share the same SDK cache used by tests. ## Limitations - Maximum of 5 buffers per external - Buffers are single-channel only. Use multiple buffers for multi-channel audio. - Max/MSP: Windows builds require Visual Studio or equivalent MSVC toolchain - AudioUnit: macOS only - CLAP: first CMake configure requires network access to fetch CLAP headers (cached afterward) - VST3: first CMake configure requires network access to fetch VST3 SDK (~50MB, cached afterward); GPL3/proprietary dual license - LV2: first CMake configure requires network access to fetch LV2 headers (cached afterward) - SuperCollider: first CMake configure requires network access to fetch SC headers (~80MB tarball, cached afterward) - VCV Rack: first build requires network access to fetch Rack SDK (cached afterward); `RACK_DIR` env var can override auto-download; per-sample `perform(n=1)` has higher CPU overhead than block-based processing - Daisy: requires `arm-none-eabi-gcc` cross-compiler; first clone of libDaisy requires network access and `git`; v1 targets Daisy Seed only (no board-specific knob/CV mapping) - Circle: requires `aarch64-none-elf-gcc` (or `arm-none-eabi-gcc` for Pi Zero) cross-compiler; first clone of Circle SDK requires network access and `git`; output-only (no audio input capture); single-plugin mode requires manual GPIO/ADC code for parameter control; multi-plugin mode (`--graph`) supports linear chains and arbitrary DAGs (fan-out, fan-in via mixer nodes) but no buffers ## Requirements ### Runtime - Python >= 3.10 - C/C++ compiler (gcc, clang) ### PureData builds - make - PureData headers (typically installed with PureData) ### Max/MSP builds - CMake >= 3.19 - git (for cloning max-sdk-base) ### ChucK builds - make - C/C++ compiler (clang on macOS, gcc on Linux) - ChucK (for running the chugin) ### AudioUnit builds - macOS (AudioUnit is macOS-only) - CMake >= 3.19 - C/C++ compiler (clang via Xcode or Command Line Tools) ### CLAP builds - CMake >= 3.19 - C/C++ compiler (clang, gcc) - Network access on first configure (to fetch CLAP headers) ### VST3 builds - CMake >= 3.19 - C/C++ compiler (clang, gcc, MSVC) - Network access on first configure (to fetch VST3 SDK, ~50MB) ### LV2 builds - CMake >= 3.19 - C/C++ compiler (clang, gcc) - Network access on first configure (to fetch LV2 headers) ### SuperCollider builds - CMake >= 3.19 - C/C++ compiler (clang, gcc) - Network access on first configure (to fetch SC plugin headers) ### VCV Rack builds - make - C/C++ compiler (clang, gcc) - Network access on first build (Rack SDK auto-downloaded and cached; override with `RACK_DIR` env var) ### Daisy builds - make - `arm-none-eabi-gcc` ([ARM GNU Toolchain Downloads](https://developer.arm.com/downloads/-/arm-gnu-toolchain-downloads) -- select `arm-none-eabi`) - git (for cloning libDaisy on first build) - Network access on first build (to clone libDaisy + submodules) ### Circle builds - make - `aarch64-none-elf-gcc` ([ARM GNU Toolchain Downloads](https://developer.arm.com/downloads/-/arm-gnu-toolchain-downloads) -- select `aarch64-none-elf`) or `arm-none-eabi-gcc` (for Pi Zero) - git (for cloning Circle SDK on first build) - Network access on first build (to clone Circle) ### macOS Install Xcode or Command Line Tools: ```bash xcode-select --install ``` ### Linux / Organelle Standard build tools (gcc, make) are typically pre-installed. ## Cross-Compilation Note Build artifacts are platform-specific. When moving projects between macOS and Linux/Organelle: ```bash make clean make all ``` ## Development ```bash git clone https://github.com/samesimilar/gen-dsp.git cd gen-dsp uv venv && uv pip install -e ".[dev]" source .venv/bin/activate make test ``` ### Building Example Plugins The Makefile includes targets for generating and building example plugins from the test fixtures: ```bash make example-pd # PureData external make example-max # Max/MSP external make example-chuck # ChucK chugin make example-au # AudioUnit plugin (macOS only) make example-clap # CLAP plugin make example-vst3 # VST3 plugin make example-lv2 # LV2 plugin make example-sc # SuperCollider UGen make example-vcvrack # VCV Rack module (auto-downloads Rack SDK) make example-daisy # Daisy firmware (requires arm-none-eabi-gcc) make example-circle # Circle kernel image (requires aarch64-none-elf-gcc) make examples # All platforms ``` Override the fixture, name, or buffers: ```bash make example-chuck FIXTURE=RamplePlayer NAME=rampleplayer BUFFERS="--buffers sample" ``` Available fixtures: `gigaverb` (default, no buffers), `RamplePlayer` (has buffers), `spectraldelayfb`. Output goes to `build/examples/`. ### Adding New Backends gen-dsp uses a platform registry system that makes it straightforward to add support for new audio platforms (Bela, Daisy, etc.). See [docs/new_backends.md](docs/new_backends.md) for a complete guide. ## Attribution The gen~ language was created by [Graham Wakefield](https://github.com/grrrwaaa) at Cycling '74. This project builds on the original idea and work of [gen_ext](https://github.com/samesimilar/gen_ext) by Michael Spears. Test fixtures include code exported from examples bundled with Max: - gigaverb: ported from Juhana Sadeharju's implementation - spectraldelayfb: from gen~.spectraldelay_feedback The Daisy backend was informed by techniques from [oopsy](https://github.com/electro-smith/oopsy) by Electrosmith and contributors, including Graham Wakefield. The Circle backend uses [Circle](https://github.com/rsta2/circle) by Rene Stange, a C++ bare metal programming environment for the Raspberry Pi. ## License MIT License. See [LICENSE](LICENSE) for details. Note: Generated gen~ code is subject to Cycling '74's license terms.
text/markdown
Michael Spears, Shakeeb Alireza
Shakeeb Alireza <shakfu@users.noreply.github.com>
null
null
MIT
puredata, pd, max, msp, gen~, dsp, audio, external
[ "Development Status :: 3 - Alpha", "Environment :: Console", "Intended Audience :: Developers", "Intended Audience :: Other Audience", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Topic :: Multimedia :: Sound/Audio", "Topic :: Software Development :: Code Generators" ]
[]
null
null
>=3.10
[]
[]
[]
[]
[]
[]
[]
[ "Homepage, https://github.com/shakfu/gen-dsp", "Repository, https://github.com/shakfu/gen-dsp", "Changelog, https://github.com/shakfu/gen-dsp/blob/master/CHANGELOG.md", "Issues, https://github.com/shakfu/gen-dsp/issues" ]
twine/6.2.0 CPython/3.13.2
2026-02-21T01:11:29.772442
gen_dsp-0.1.12.tar.gz
186,376
66/54/0d56e48fe1e1c1a8fb36c1daa11a88f6879b5467fd392f46aba07446c0e3/gen_dsp-0.1.12.tar.gz
source
sdist
null
false
762161ea1b0e27db3e98c13b021dfbc2
8a932e8647e4cd13b32183283d38f88cb7561f9994139d0d756d36e0de6aa6ad
66540d56e48fe1e1c1a8fb36c1daa11a88f6879b5467fd392f46aba07446c0e3
null
[]
240
2.4
nirs4all
0.8.0
A comprehensive Python library for Near-Infrared Spectroscopy (NIRS) data analysis with ML/DL pipelines.
<div align="center"> <img src="docs/source/assets/nirs4all_logo.png" width="300" alt="NIRS4ALL Logo"> <img src="docs/source/assets/logo-cirad-en.jpg" width="300" alt="CIRAD Logo"> # NIRS4ALL **A comprehensive Python library for Near-Infrared Spectroscopy data analysis** [![PyPI version](https://img.shields.io/pypi/v/nirs4all.svg)](https://pypi.org/project/nirs4all/) [![Python 3.11+](https://img.shields.io/badge/python-3.11+-blue.svg)](https://www.python.org/downloads/) [![License: CeCILL-2.1](https://img.shields.io/badge/license-CeCILL--2.1-blue.svg)](LICENSE) [![Code style: ruff](https://img.shields.io/badge/code%20style-ruff-000000.svg)](https://github.com/astral-sh/ruff) [Documentation](https://nirs4all.readthedocs.io) • [Installation](#installation) • [Quick Start](#quick-start) • [Examples](examples/) • [Contributing](CONTRIBUTING.md) </div> --- ## Overview NIRS4ALL bridges the gap between spectroscopic data and machine learning by providing a unified framework for **data loading**, **preprocessing**, **model training**, and **evaluation**. Built for researchers and practitioners working with Near-Infrared Spectroscopy data. <div align="center"> <img src="docs/source/assets/pipeline.jpg" width="400" alt="Performance Heatmap"> </div> ### Key Features - **NIRS-Specific Preprocessing** — SNV, MSC, Savitzky-Golay, Norris-Williams, wavelet denoise, OSC/EPO, and 30+ spectral transforms - **Advanced PLS Models** — AOM-PLS, POP-PLS, OPLS, DiPLS, MBPLS, and 15+ PLS variants with automatic operator selection - **Multi-Backend ML** — Seamless integration with scikit-learn, TensorFlow, PyTorch, and JAX - **Declarative Pipelines** — Define complex workflows with simple, readable syntax - **Parallel Execution** — Multi-core pipeline variant execution via joblib - **Hyperparameter Tuning** — Built-in Optuna integration for automated optimization - **Rich Visualizations** — Performance heatmaps, candlestick plots, SHAP explanations - **Model Deployment** — Export trained pipelines as portable `.n4a` bundles - **sklearn Compatible** — `NIRSPipeline` wrapper for SHAP, cross-validation, and more <div align="center"> <img src="docs/source/assets/heatmap.png" width="400" alt="Performance Heatmap"> <img src="docs/source/assets/candlestick.png" width="400" alt="Performance Distribution"> <img src="docs/source/assets/stacking.png" width="400" alt="Regression Scatter Plot"> <br><em>Advanced visualization capabilities for model performance analysis</em> </div> --- ## Installation ### Basic Installation ```bash pip install nirs4all ``` This installs the core library with scikit-learn support. Deep learning frameworks are optional. ### With ML Backends ```bash # TensorFlow pip install nirs4all[tensorflow] # PyTorch pip install nirs4all[torch] # JAX pip install nirs4all[jax] # All frameworks pip install nirs4all[all] # All frameworks with GPU support pip install nirs4all[all-gpu] ``` ### Conda ```bash conda install -c conda-forge nirs4all ``` ### Docker ```bash docker pull ghcr.io/gbeurier/nirs4all:latest docker run -v $(pwd):/workspace ghcr.io/gbeurier/nirs4all python my_script.py ``` ### Development Installation ```bash git clone https://github.com/GBeurier/nirs4all.git cd nirs4all pip install -e ".[dev]" ``` ### Verify Installation ```bash nirs4all --test-install # Check dependencies nirs4all --test-integration # Run integration tests nirs4all --version # Check version ``` --- ## Quick Start ### Simple API (Recommended) ```python import nirs4all from sklearn.preprocessing import MinMaxScaler from sklearn.model_selection import ShuffleSplit from sklearn.cross_decomposition import PLSRegression # Define your pipeline pipeline = [ MinMaxScaler(), {"y_processing": MinMaxScaler()}, ShuffleSplit(n_splits=3, test_size=0.25), {"model": PLSRegression(n_components=10)} ] # Train and evaluate result = nirs4all.run( pipeline=pipeline, dataset="path/to/your/data", name="MyPipeline", verbose=1 ) # Access results print(f"Best RMSE: {result.best_rmse:.4f}") print(f"Best R²: {result.best_r2:.4f}") # Export for deployment result.export("exports/best_model.n4a") ``` ### Session for Multiple Runs ```python import nirs4all from sklearn.preprocessing import MinMaxScaler from sklearn.cross_decomposition import PLSRegression from sklearn.ensemble import RandomForestRegressor with nirs4all.session(verbose=1, save_artifacts=True) as s: # Compare models with shared configuration pls_result = nirs4all.run( pipeline=[MinMaxScaler(), PLSRegression(n_components=10)], dataset="data/wheat.csv", name="PLS", session=s ) rf_result = nirs4all.run( pipeline=[MinMaxScaler(), RandomForestRegressor(n_estimators=100)], dataset="data/wheat.csv", name="RandomForest", session=s ) print(f"PLS: {pls_result.best_rmse:.4f} | RF: {rf_result.best_rmse:.4f}") ``` ### sklearn Integration with SHAP ```python import nirs4all from nirs4all.sklearn import NIRSPipeline import shap # Train with nirs4all result = nirs4all.run(pipeline, dataset) # Wrap for sklearn compatibility pipe = NIRSPipeline.from_result(result) # Use with SHAP explainer = shap.Explainer(pipe.predict, X_background) shap_values = explainer(X_test) shap.summary_plot(shap_values) ``` --- ## Pipeline Syntax NIRS4ALL uses a declarative syntax for defining pipelines: ```python from nirs4all.operators.transforms import SNV, SavitzkyGolay, FirstDerivative pipeline = [ # Preprocessing MinMaxScaler(), SNV(), SavitzkyGolay(window_length=11, polyorder=2), # Target scaling {"y_processing": MinMaxScaler()}, # Cross-validation ShuffleSplit(n_splits=5, test_size=0.2), # Models to compare {"model": PLSRegression(n_components=10)}, {"model": RandomForestRegressor(n_estimators=100)}, # Neural network with training parameters { "model": nicon, "name": "NICON-CNN", "train_params": {"epochs": 100, "patience": 20} } ] ``` ### Advanced Features ```python # Feature augmentation - generate preprocessing combinations { "feature_augmentation": { "_or_": [SNV, FirstDerivative, SavitzkyGolay], "size": [1, (1, 2)], "count": 5 } } # Hyperparameter optimization { "model": PLSRegression(), "finetune_params": { "n_trials": 50, "model_params": {"n_components": ("int", 1, 30)} } } # Branching for parallel preprocessing paths { "branch": [ [SNV(), PLSRegression(n_components=10)], [MSC(), RandomForestRegressor()] ] } # Merge branch outputs (stacking) {"merge": "predictions"} ``` --- ## Available Transforms ### NIRS-Specific Preprocessing | Transform | Description | |-----------|-------------| | `SNV` / `StandardNormalVariate` | Standard Normal Variate normalization | | `RNV` / `RobustStandardNormalVariate` | Robust Normal Variate (outlier-resistant) | | `MSC` / `MultiplicativeScatterCorrection` | Multiplicative Scatter Correction | | `SavitzkyGolay` | Smoothing and derivative computation | | `FirstDerivative` / `SecondDerivative` | Spectral derivatives | | `NorrisWilliams` | Gap derivative with segment smoothing | | `WaveletDenoise` | Multi-level wavelet denoising with thresholding | | `OSC` | Orthogonal Signal Correction (DOSC) | | `EPO` | External Parameter Orthogonalization | | `Detrend` | Remove linear/polynomial trends | | `Gaussian` | Gaussian smoothing | | `Haar` | Haar wavelet decomposition | ### Signal Processing | Transform | Description | |-----------|-------------| | `Baseline` | Baseline correction (ALS, AirPLS, ArPLS, IModPoly, SNIP, etc.) | | `ReflectanceToAbsorbance` | Convert R to A using Beer-Lambert | | `ToAbsorbance` / `FromAbsorbance` | Signal type conversion | | `KubelkaMunk` | Kubelka-Munk transform | | `Resampler` | Wavelength interpolation | | `CARS` / `MCUVE` | Feature selection methods | ### Built-in NIRS Models | Model | Description | |-------|-------------| | `AOMPLSRegressor` / `AOMPLSClassifier` | Adaptive Operator-Mixture PLS — auto-selects best preprocessing | | `POPPLSRegressor` / `POPPLSClassifier` | Per-Operator-Per-component PLS via PRESS | | `PLSDA` | PLS Discriminant Analysis | | `OPLS` / `OPLSDA` | Orthogonal PLS | | `MBPLS` | Multi-Block PLS | | `DiPLS` | Domain-Invariant PLS | | `IKPLS` | Improved Kernel PLS | | `FCKPLS` | Fractional Convolution Kernel PLS | ### Splitting Methods | Splitter | Description | |----------|-------------| | `KennardStoneSplitter` | Kennard-Stone algorithm | | `SPXYSplitter` | Sample set Partitioning based on X and Y | | `SPXYFold` / `SPXYGFold` | SPXY-based K-Fold cross-validation (with group support) | | `KMeansSplitter` | K-means clustering based split | | `KBinsStratifiedSplitter` | Binned stratification for continuous targets | See [Preprocessing Guide](docs/source/user_guide/preprocessing/) for complete reference. --- ## Examples The `examples/` directory is organized by topic: ### User Examples (`examples/user/`) | Category | Examples | |----------|----------| | **Getting Started** | Hello world, basic regression, classification, visualization | | **Data Handling** | Multi-source, data loading, metadata | | **Preprocessing** | SNV, MSC, derivatives, custom transforms | | **Models** | Multi-model, hyperparameter tuning, stacking, PLS variants | | **Cross-Validation** | KFold, group splits, nested CV | | **Deployment** | Export, prediction, workspace management | | **Explainability** | SHAP basics, sklearn integration, feature selection | ### Reference Examples (`examples/reference/`) Complete syntax reference and advanced pipeline patterns. Run examples: ```bash cd examples ./run.sh # Run all ./run.sh -i 1 # Run by index ./run.sh -n "U01*" # Run by pattern ``` --- ## Documentation | Section | Description | |---------|-------------| | [**User Guide**](docs/user_guide/) | Preprocessing, API migration, augmentation | | [**API Reference**](docs/api/) | Module-level API, sklearn integration, data handling | | [**Specifications**](docs/specifications/) | Pipeline syntax, config format, metrics | | [**Explanations**](docs/explanations/) | SHAP, resampling, SNV theory | Full documentation: [nirs4all.readthedocs.io](https://nirs4all.readthedocs.io) --- ## Research Applications NIRS4ALL has been used in published research: > **Houngbo, M. E., et al. (2024).** *Convolutional neural network allows amylose content prediction in yam (Dioscorea alata L.) flour using near infrared spectroscopy.* **Journal of the Science of Food and Agriculture, 104(8), 4915-4921.** John Wiley & Sons, Ltd. --- ## Citation If you use NIRS4ALL in your research, please cite: ```bibtex @software{beurier2025nirs4all, author = {Gregory Beurier and Denis Cornet and Lauriane Rouan}, title = {NIRS4ALL: Open spectroscopy for everyone}, url = {https://github.com/GBeurier/nirs4all}, version = {0.7.1}, year = {2026}, } ``` --- ## Contributing We welcome contributions! Please see [CONTRIBUTING.md](CONTRIBUTING.md) for guidelines. - [Report bugs](https://github.com/GBeurier/nirs4all/issues) - [Request features](https://github.com/GBeurier/nirs4all/issues) - [Improve documentation](docs/) --- ## License This project is licensed under the [CeCILL-2.1 License](LICENSE) — a French free software license compatible with GPL. --- ## Acknowledgments - [CIRAD](https://www.cirad.fr/) for supporting this research - The open-source scientific Python community <div align="center"> <br> <strong>Made for the spectroscopy community</strong> </div>
text/markdown
null
Gregory Beurier <beurier@cirad.fr>, Denis Cornet <denis.cornet@cirad.fr>, Lauriane Rouan <lauriane.rouan@cirad.fr>
null
Gregory Beurier <gregory.beurier@cirad.fr>
This repository (project **nirs4all**) is distributed under a **DUAL LICENSE**. 1) **Open-source (default)**: **AGPL-3.0-or-later** - See the full text in `LICENSES/AGPL-3.0-or-later.txt`. - Default SPDX: **AGPL-3.0-or-later** (add it to file headers). - Optional variants: **GPL-3.0-or-later** or **CeCILL-2.1**. 2) **Commercial**: - A proprietary license allowing closed-source use, including SaaS, without the obligation to publish your code. - See `LICENSES/COMMERCIAL-LICENSE.md` for terms and how to obtain it. - Contact: <denis.cornet@cirad.fr>. Unless a commercial agreement is explicitly in place, this code is governed by **AGPL-3.0-or-later**. Recommended SPDX header for Python files: # SPDX-License-Identifier: AGPL-3.0-or-later
nirs, near-infrared, spectroscopy, chemometrics, machine-learning, deep-learning, preprocessing, pls, sklearn, tensorflow, pytorch, jax
[ "Development Status :: 4 - Beta", "Intended Audience :: Science/Research", "Intended Audience :: Developers", "License :: OSI Approved :: CEA CNRS Inria Logiciel Libre License, version 2.1 (CeCILL-2.1)", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Scientific/Engineering", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Scientific/Engineering :: Chemistry", "Typing :: Typed" ]
[]
null
null
>=3.11
[]
[]
[]
[ "numpy>=1.24.0", "pandas>=2.0.0", "scipy>=1.10.0", "scikit-learn>=1.2.0", "polars>=1.0.0", "pyarrow>=14.0.0", "duckdb>=1.0.0", "h5py>=3.8.0", "umap-learn", "PyWavelets>=1.4.0", "pybaselines>=1.2.1", "optuna>=3.0.0", "matplotlib>=3.7.0", "seaborn>=0.12.0", "shap>=0.42.0", "joblib>=1.2.0", "jsonschema>=4.17.0", "pyyaml>=6.0.0", "packaging>=23.0", "pydantic>=2.0.0", "kennard-stone>=2.2.0", "ikpls>=1.1.0", "pyopls>=20.0", "trendfitter>=0.0.6", "tensorflow>=2.14.0; extra == \"tensorflow\"", "tensorflow[and-cuda]>=2.14.0; extra == \"gpu\"", "torch>=2.1.0; extra == \"torch\"", "tabpfn>=2.0.0; extra == \"torch\"", "keras>=3.0.0; extra == \"keras\"", "jax>=0.4.20; extra == \"jax\"", "jaxlib>=0.4.20; extra == \"jax\"", "flax>=0.8.0; extra == \"jax\"", "autogluon>=1.0.0; extra == \"autogluon\"", "tensorflow>=2.14.0; extra == \"all\"", "torch>=2.1.0; extra == \"all\"", "keras>=3.0.0; extra == \"all\"", "jax>=0.4.20; extra == \"all\"", "jaxlib>=0.4.20; extra == \"all\"", "flax>=0.8.0; extra == \"all\"", "tensorflow[and-cuda]>=2.14.0; extra == \"all-gpu\"", "torch>=2.1.0; extra == \"all-gpu\"", "keras>=3.0.0; extra == \"all-gpu\"", "jax[cuda12]>=0.4.20; extra == \"all-gpu\"", "jaxlib>=0.4.20; extra == \"all-gpu\"", "flax>=0.8.0; extra == \"all-gpu\"", "pytest>=7.4.0; extra == \"dev\"", "pytest-cov>=4.1.0; extra == \"dev\"", "pytest-xdist>=3.5.0; extra == \"dev\"", "pytest-timeout>=2.2.0; extra == \"dev\"", "pytest-benchmark>=4.0.0; extra == \"dev\"", "build>=1.0.0; extra == \"dev\"", "ruff>=0.1.0; extra == \"dev\"", "mypy>=1.7.0; extra == \"dev\"", "sphinx>=7.0.0; extra == \"docs\"", "sphinx-rtd-theme>=2.0.0; extra == \"docs\"", "myst-parser>=2.0.0; extra == \"docs\"", "sphinx-copybutton>=0.5.0; extra == \"docs\"", "sphinx-design>=0.5.0; extra == \"docs\"", "sphinxcontrib-mermaid; extra == \"docs\"", "pypandoc>=1.12; extra == \"reports\"", "PyPDF2>=3.0.0; extra == \"reports\"", "pdf2image>=1.16.0; extra == \"reports\"" ]
[]
[]
[]
[ "Homepage, https://github.com/GBeurier/nirs4all", "Documentation, https://nirs4all.readthedocs.io", "Repository, https://github.com/GBeurier/nirs4all", "Issues, https://github.com/GBeurier/nirs4all/issues", "Changelog, https://github.com/GBeurier/nirs4all/blob/main/CHANGELOG.md" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T01:11:14.067035
nirs4all-0.8.0.tar.gz
1,587,392
30/aa/e41aabe196015f1464d72dbb3ffb8c5ab54b96824321d2ebd92c2d149a2d/nirs4all-0.8.0.tar.gz
source
sdist
null
false
8b7fe438b13b6eaa9a3f15c295093b6e
ff99ea2509f6be8007b8155746189e28286f0d6b4b2443eb552947532a9b0499
30aae41aabe196015f1464d72dbb3ffb8c5ab54b96824321d2ebd92c2d149a2d
null
[ "LICENSE", "LICENSE_FR" ]
225
2.4
pytest-data-loader
0.7.2
Pytest plugin for loading test data for data-driven testing (DDT)
pytest-data-loader ====================== [![PyPI](https://img.shields.io/pypi/v/pytest-data-loader)](https://pypi.org/project/pytest-data-loader/) [![Supported Python versions](https://img.shields.io/pypi/pyversions/pytest-data-loader.svg)](https://pypi.org/project/pytest-data-loader/) [![test](https://github.com/yugokato/pytest-data-loader/actions/workflows/test.yml/badge.svg?branch=main)](https://github.com/yugokato/pytest-data-loader/actions/workflows/test.yml?query=branch%3Amain) [![Code style ruff](https://img.shields.io/badge/code%20style-ruff-000000.svg)](https://docs.astral.sh/ruff/) `pytest-data-loader` is a `pytest` plugin that simplifies loading test data from files for data-driven testing. It supports not only loading a single file, but also dynamic test parametrization either by splitting file data into parts or by loading multiple files from a directory. ## Installation ```bash pip install pytest-data-loader ``` ## Quick Start ```python from pytest_data_loader import load @load("data", "example.json") def test_example(data): """ Loads data/example.json and injects it as the "data" fixture. example.json: '{"foo": 1, "bar": 2}' """ assert "foo" in data ``` ## Usage The plugin provides three data loaders — `@load`, `@parametrize`, and `@parametrize_dir` — available as decorators for loading test data. Each loader takes two positional arguments: - `fixture_names`: Name(s) of the fixture(s) that will be made available to the test function. It supports either one (receiving file data) or two (receiving file path and file data) fixture names - `path`: Path to the file or directory to load data from. It can be either an absolute path or a path relative to one of the project's data directories. When a relative path is provided, the plugin searches upward from the test file's directory toward the Pytest root directory to find the nearest data directory containing the target file. > [!TIP] > - By default, the plugin looks for a directory named `data` when resolving relative paths. This default name can be > customized using an INI option. See the [INI Options](#ini-options) section for details > - Each data loader supports different optional keyword arguments to customize how the data is loaded. See the > [Loader Options](#loader-options) section for details ## Examples Given you have the following project structure: ``` .(pytest rootdir) ├── data/ │ ├── data1.json │ ├── data2.txt │ └── images/ │ ├── image.gif │ ├── image.jpg │ └── image.png ├── tests1/ │ └── test_something.py └── tests2/ ├── data/ │ ├── data1.txt │ └── data2.txt └── test_something_else.py ``` ### 1. Load file data — `@load` `@load` is a file loader that loads the file content and passes it to the test function. ```python # test_something.py from pytest_data_loader import load @load("data", "data1.json") def test_something1(data): """ data1.json: '{"foo": 1, "bar": 2}' """ assert data == {"foo": 1, "bar": 2} @load(("file_path", "data"), "data2.txt") def test_something2(file_path, data): """ data2.txt: "line1\nline2\nline3" """ assert file_path.name == "data2.txt" assert data == "line1\nline2\nline3" ``` ```shell $ pytest tests1/test_something.py -v ================================ test session starts ================================= <snip> collected 2 items tests1/test_something.py::test_something1[data1.json] PASSED [ 50%] tests1/test_something.py::test_something2[data2.txt] PASSED [100%] ================================= 2 passed in 0.01s ================================== ``` > [!NOTE] > If both `./tests1/test_something.py` and `./tests2/test_something_else.py` happen to have the above same loader > definitions, the first test function will load `./data/data1.json` for both test files, and the second test function > will load `data2.txt` from each test file's **nearest** `data` directory. This ensures that each test file loads data > from its nearest data directory. > This behavior applies to all loaders. ### 2. Parametrize file data — `@parametrize` `@parametrize` is a file loader that dynamically parametrizes the decorated test function by splitting the loaded file content into logical parts. The test function will then receive the part data as loaded data for the current test. ```python # test_something.py from pytest_data_loader import parametrize @parametrize("data", "data1.json") def test_something1(data): """ data1.json: '{"foo": 1, "bar": 2}' """ assert data in [("foo", 1), ("bar", 2)] @parametrize(("file_path", "data"), "data2.txt") def test_something2(file_path, data): """ data2.txt: "line1\nline2\nline3" """ assert file_path.name == "data2.txt" assert data in ["line1", "line2", "line3"] ``` ```shell $ pytest tests1/test_something.py -v ================================ test session starts ================================= <snip> collected 5 items tests1/test_something.py::test_something1[data1.json:part1] PASSED [ 20%] tests1/test_something.py::test_something1[data1.json:part2] PASSED [ 40%] tests1/test_something.py::test_something2[data2.txt:part1] PASSED [ 60%] tests1/test_something.py::test_something2[data2.txt:part2] PASSED [ 80%] tests1/test_something.py::test_something2[data2.txt:part3] PASSED [100%] ================================= 5 passed in 0.01s ================================== ``` > [!TIP] > - You can apply your own logic by specifying the `parametrizer_func` loader option > - By default, the plugin will apply the following logic for splitting file content: > - Text file: Each line in the file > - JSON file: > - object: Each key–value pair in the object > - array: Each item in the array > - other types (string, number, boolean, null): The whole content as single data > - Binary file: Unsupported. Requires specifying a custom split logic as the `parametrizer_func` loader option ### 3. Parametrize files in a directory — `@parametrize_dir` `@parametrize_dir` is a directory loader that dynamically parametrizes the decorated test function with the contents of the files stored in the specified directory. The test function will then receive the content of each file as loaded data for the current test. ```python # test_something.py from pytest_data_loader import parametrize_dir @parametrize_dir("data", "images") def test_something(data): """ images dir: contains 3 image files """ assert isinstance(data, bytes) ``` ```shell $ pytest tests1/test_something.py -v ================================ test session starts ================================= <snip> collected 3 items tests1/test_something.py::test_something[images/image.gif] PASSED [ 33%] tests1/test_something.py::test_something[images/image.jpg] PASSED [ 66%] tests1/test_something.py::test_something[images/image.png] PASSED [100%] ================================= 3 passed in 0.01s ================================== ``` > [!NOTE] > - File names starting with a dot (.) are considered hidden files regardless of your platform. > These files are automatically excluded from the parametrization. > - Specify `recursive=True` to include files in subdirectories ## Lazy Loading Lazy loading is enabled by default for all data loaders to improve efficiency, especially with large datasets. During test collection, pytest receives a lazy object as a test parameter instead of the actual data. The data is resolved only when it is needed during test setup. If you need to disable this behavior for a specific test, pass `lazy_loading=False` to the data loader. > [!NOTE] > Lazy loading for the `@parametrize` loader works slightly differently from other loaders. Since Pytest needs to know > the total number of parameters in advance, the plugin still needs to inspect the file data and split it once during > test collection phase. But once it's done, those part data will not be kept as parameter values and will be loaded > lazily later. ## File Reader You can specify **any file reader** that accepts a file-like object returned by `open()`. This includes built-in readers, third-party library readers, and your own custom readers. File read options (e.g., `mode`, `encoding`, etc.) can also be provided and will be passed to `open()`. This can be done either as a `conftest.py` level registration or as a test-level configuration. If both are done, the test level configuration takes precedence over `conftest.py` level registration. If multiple `conftest.py` files register a reader for the same file extension, the closest one from the current test becomes effective. Below are some common examples of file readers you might use: | File type | Examples | Notes | |-----------|---------------------------------------------------|--------------------------------------------------| | .json | `json.load` | The plugin automatically uses this by default | | .csv | `csv.reader`, `csv.DictReader`, `pandas.read_csv` | `pandas.read_csv` requires `pandas` | | .yml | `yaml.safe_load`, `yaml.safe_load_all` | Requires `PyYAML` | | .xml | `xml.etree.ElementTree.parse` | | | .toml | `tomllib.load` | `tomli.load` for Python <3.11 (Requires `tomli`) | | .ini | `configparser.ConfigParser().read_file` | | | .pdf | `pypdf.PdfReader` | Requires `pypdf` | Here are some examples of loading a CSV file using the built-in CSV readers with file read options: ### 1. `conftest.py` level registration Register a file reader using `pytest_data_loader.register_reader()`. It takes a file extension and a file reader as positional arguments, and file read options as keyword arguments. ```python # conftest.py import csv import pytest_data_loader pytest_data_loader.register_reader(".csv", csv.reader, newline="") ``` The registered file reader automatically applies to all tests located in the same directory and any of its subdirectories. ```python # test_something.py from pytest_data_loader import load @load("data", "data.csv") def test_something(data): """Load CSV file with registered file reader""" for row in data: assert isinstance(row, list) ``` ### 2. Per-test configuration with loader options Specify a file reader with the `file_reader` loader option. This applies only to the configured test, and overrides the one registered in `conftest.py`. ```python # test_something.py import csv from pytest_data_loader import load, parametrize @load("data", "data.csv", file_reader=csv.reader, encoding="utf-8-sig", newline="") def test_something1(data): """Load CSV file with csv.reader reader""" for row in data: assert isinstance(row, list) @parametrize("data", "data.csv", file_reader=csv.DictReader, encoding="utf-8-sig", newline="") def test_something2(data): """Parametrize CSV file data with csv.DictReader reader""" assert isinstance(data, dict) ``` > [!NOTE] > If only read options are specified without a `file_reader` in a loader, the plugin will search for an existing file > reader registered in `conftest.py` if there is any, and applies it with the new read options for the test. But if > only a `file_reader` is specified with no read options in a loader, no read options will be applied. > [!TIP] > - A file reader must take one argument (a file-like object returned by `open()`) > - If you need to pass options to the file reader, use `lambda` function or a regular function. > eg. `file_reader=lambda f: csv.reader(f, delimiter=";")` > - You can adjust the final data the test function receives using loader functions. For example, > the following code will parametrize the test with the text data from each PDF page > ```python > @parametrize( > "data", > "test.pdf", > file_reader=pypdf.PdfReader, > parametrizer_func=lambda r: r.pages, > process_func=lambda p: p.extract_text().rstrip(), > mode="rb" > ) > def test_something(data: str): > ... > ``` ## Loader Options Each loader supports different optional parameters you can use to change how your data is loaded. ### @load - `lazy_loading`: Enable or disable lazy loading - `file_reader`: A file reader the plugin should use to read the file data - `onload_func`: A function to transform or preprocess loaded data before passing it to the test function - `id`: The parameter ID for the loaded data. If not specified, the relative or absolute file path is used - `**read_options`: File read options the plugin passes to `open()`. Supports only `mode`, `encoding`, `errors`, and `newline` options > [!NOTE] > `onload_func` must take either one (data) or two (file path, data) arguments. When `file_reader` is provided, the data is the reader object itself. ### @parametrize - `lazy_loading`: Enable or disable lazy loading - `file_reader`: A file reader the plugin should use to read the file data - `onload_func`: A function to adjust the shape of the loaded data before splitting into parts - `parametrizer_func`: A function to customize how the loaded data should be split - `filter_func`: A function to filter the split data parts. Only matching parts are included as test parameters - `process_func`: A function to adjust the shape of each split data before passing it to the test function - `marker_func`: A function to apply Pytest marks to matching part data - `id_func`: A function to generate a parameter ID for each part data - `**read_options`: File read options the plugin passes to `open()`. Supports only `mode`, `encoding`, `errors`, and `newline` options > [!NOTE] > Each loader function must take either one (data) or two (file path, data) arguments. When `file_reader` is provided, the data is the reader object itself. ### @parametrize_dir - `lazy_loading`: Enable or disable lazy loading - `recursive`: Recursively load files from all subdirectories of the given directory. Defaults to `False` - `file_reader_func`: A function to specify file readers to matching file paths - `filter_func`: A function to filter file paths. Only the contents of matching file paths are included as the test parameters - `process_func`: A function to adjust the shape of each loaded file's data before passing it to the test function - `marker_func`: A function to apply Pytest marks to matching file paths - `read_option_func`: A function to specify file read options the plugin passes to `open()` to matching file paths. Supports only `mode`, `encoding`, `errors`, and `newline` options. It must return these options as a dictionary > [!NOTE] > - `process_func` must take either one (data) or two (file path, data) arguments > - `file_reader_func`, `filter_func`, `marker_func`, and `read_option_func` must take only one argument (file path) ## INI Options ### `data_loader_dir_name` The base directory name to load test data from. When a relative file or directory path is provided to a data loader, it is resolved relative to the nearest matching data directory in the directory tree. Plugin default: `data` ### `data_loader_root_dir` Absolute or relative path to the project's root directory. By default, the search is limited to within pytest's rootdir, which may differ from the project's top-level directory. Setting this option allows data directories located outside pytest's rootdir to be found. Environment variables are supported using the `${VAR}` or `$VAR` (or `%VAR%` for windows) syntax. Plugin default: Pytest rootdir (`config.rootpath`) ### `data_loader_strip_trailing_whitespace` Automatically remove trailing whitespace characters when loading text data. Plugin default: `true`
text/markdown
Yugo Kato
null
null
null
null
pytest, data-driven testing, DDT
[ "Development Status :: 4 - Beta", "Framework :: Pytest", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Topic :: Software Development :: Quality Assurance", "Topic :: Software Development :: Testing" ]
[]
null
null
>=3.10
[]
[]
[]
[ "pytest<10,>=7.0.0", "pluggy>=1.2.0", "typing_extensions>=4.1.0; python_version < \"3.11\"", "mypy<2,>=1.15.0; extra == \"lint\"", "pre-commit<5,>=3.0.0; extra == \"lint\"", "ruff<0.16.0,>=0.15.0; extra == \"lint\"", "pytest-mock<4,>=3.0.0; extra == \"test\"", "pytest-smoke; extra == \"test\"", "pytest-xdist[psutil]<4,>=2.3.0; extra == \"test\"", "pandas>=2.0.0; extra == \"test\"", "pypdf>=6.1.0; extra == \"test\"", "pyyaml>=6.0.0; extra == \"test\"", "tomli>=2.3.0; python_version < \"3.11\" and extra == \"test\"", "tox<5,>=4.0.0; extra == \"dev\"", "tox-uv<2,>=1.0.0; extra == \"dev\"", "pytest-data-loader[lint,test]; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://github.com/yugokato/pytest-data-loader" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T01:09:58.782267
pytest_data_loader-0.7.2.tar.gz
68,960
b5/b1/6aba6445022ba44c965520a1a165e4102051e15ba536e9bdbb117722e4dc/pytest_data_loader-0.7.2.tar.gz
source
sdist
null
false
dae5f1445a97bf9a9f59841c86a76c69
8e0023a632655af3abb790ed62883468fb8d2bc3b5542eef6cb66022745dc6cb
b5b16aba6445022ba44c965520a1a165e4102051e15ba536e9bdbb117722e4dc
null
[]
224
2.4
raster-tools
0.9.11
Tools for processing geospatial data
## Raster Tools -------------- Python tools for lazy, parallel geospatial raster processing. ## Introduction --------------- Raster Tools is a Python package that facilitates a wide range of geospatial, statistical, and machine learning analyses using delayed and automated parallel processing for rasters. It seeks to bridge the gaps in Python's data stack for processing raster data and to make building processing and analysis pipelines easier. With Raster Tools, operations can be easily chained together, eliminating the need to write intermediate results and saving on storage space. The use of Dask, also enables [out-of-core processing](https://en.wikipedia.org/wiki/External_memory_algorithm) so rasters larger than the available memory can be processed in chunks. Under the hood, Raster Tools uses Dask to parallelize tasks and delay operations, [Rasterio](https://github.com/rasterio/rasterio), [rioxarray](https://github.com/corteva/rioxarray), and [odc-geo](https://github.com/opendatacube/odc-geo) for geospatial operations, and [Numba](https://github.com/numba/numba) for accelerating Python code. Limited support is also provided for working with Vector data using [dask-geopandas](https://github.com/geopandas/dask-geopandas). ## Install ---------- #### Pip ```sh pip install raster-tools ``` #### Conda ```sh conda install -c conda-forge cfgrib dask-geopandas dask-image fiona netcdf4 numba odc-geo pyogrio rioxarray scipy pip install --no-deps raster-tools ``` ## Helpful Links - [How to Contribute](./CONTRIBUTING.md) - [Documentation](https://um-rmrs.github.io/raster_tools/) - [Notebooks & Tutorials](./notebooks/README.md) - [PyPi link](https://pypi.org/project/raster-tools/) - [Installation](./notebooks/install_raster_tools.md) ## Package Dependencies - [cfgrib](https://github.com/ecmwf/cfgrib) - [dask](https://dask.org/) - [dask_image](https://image.dask.org/en/latest/) - [dask-geopandas](https://dask-geopandas.readthedocs.io/en/stable/) - [fiona](https://fiona.readthedocs.io/en/stable/) - [geopandas](https://geopandas.org/en/stable/) - [netcdf](https://unidata.github.io/netcdf4-python/) - [numba](https://numba.pydata.org/) - [numpy](https://numpy.org/doc/stable/) - [odc-geo](https://odc-geo.readthedocs.io/en/latest/) - [pyproj](https://pyproj4.github.io/pyproj/stable/) - [rasterio](https://rasterio.readthedocs.io/en/latest/) - [rioxarray](https://corteva.github.io/rioxarray/stable/) - [shapely 2](https://shapely.readthedocs.io/en/stable/) - [scipy](https://docs.scipy.org/doc/scipy/) - [xarray](https://xarray.pydata.org/en/stable/)
text/markdown
null
null
null
null
GPL-3.0
null
[ "Intended Audience :: Science/Research", "License :: OSI Approved :: GNU General Public License v3 (GPLv3)", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Topic :: Scientific/Engineering :: GIS" ]
[]
https://github.com/UM-RMRS/raster_tools
null
>=3.9
[]
[]
[]
[ "affine", "cfgrib", "dask", "dask-geopandas", "dask-image", "fiona", "geopandas", "netcdf4", "numba", "numpy>=1.22", "odc-geo<=0.4.11", "pandas", "pyogrio", "pyproj", "rioxarray", "rasterio", "scipy", "shapely>=2.0", "xarray" ]
[]
[]
[]
[ "Bug Tracker, https://github.com/UM-RMRS/raster_tools/issues" ]
twine/6.2.0 CPython/3.12.12
2026-02-21T01:09:37.154880
raster_tools-0.9.11.tar.gz
157,842
d8/ee/8cd3071c2ef51eccd1ad4449a2e9babf0a1bd92802b8c7f8d19978f637a0/raster_tools-0.9.11.tar.gz
source
sdist
null
false
e5fc471568cfb96e992be25f8a5d3fa0
37a9106d0993d0a118a80f4cb55cdea89bafe33b81195888a307dd27d473fd45
d8ee8cd3071c2ef51eccd1ad4449a2e9babf0a1bd92802b8c7f8d19978f637a0
null
[ "LICENSE" ]
233
2.4
comfy-test
0.2.3
Installation testing infrastructure for ComfyUI custom nodes
# comfy-test Testing infrastructure for ComfyUI custom nodes. Test your nodes install and work correctly across **Linux**, **macOS**, **Windows**, and **Windows Portable**. No pytest code needed. ## Quick Start Add these files to your custom node repository: ### 1. `comfy-test.toml` ```toml [test] # Name is auto-detected from directory [test.workflows] cpu = "all" # Run all workflows in workflows/ folder ``` ### 2. `.github/workflows/test-install.yml` ```yaml name: Test Installation on: [push, pull_request] jobs: test: uses: PozzettiAndrea/comfy-test/.github/workflows/test-matrix.yml@main ``` ### 3. `workflows/test.json` A minimal ComfyUI workflow that uses your nodes. Export from ComfyUI. **Done!** Push to GitHub and your tests will run automatically on all platforms. ## Test Levels comfy-test runs 7 test levels in sequence: | Level | Name | What It Does | |-------|------|--------------| | 1 | **SYNTAX** | Check project structure (pyproject.toml/requirements.txt), CP1252 compatibility | | 2 | **INSTALL** | Clone ComfyUI, create environment, install node + dependencies | | 3 | **REGISTRATION** | Start server, verify nodes appear in `/object_info` | | 4 | **INSTANTIATION** | Test each node's constructor | | 5 | **STATIC_CAPTURE** | Screenshot workflows (no execution) | | 6 | **VALIDATION** | 4-level workflow validation (schema, graph, introspection, partial execution) | | 7 | **EXECUTION** | Run workflows end-to-end, capture outputs | Each level depends on previous levels. You can run up to a specific level with `--level`: ```bash comfy-test run --level registration # Runs: SYNTAX -> INSTALL -> REGISTRATION ``` ## Workflow Validation (4 Levels) The VALIDATION level runs comprehensive checks before execution: | Level | Name | What It Checks | |-------|------|----------------| | 1 | **Schema** | Widget values match allowed enums, types, and ranges | | 2 | **Graph** | Connections are valid, all referenced nodes exist | | 3 | **Introspection** | Node definitions are well-formed (INPUT_TYPES, RETURN_TYPES, FUNCTION) | | 4 | **Partial Execution** | Runs non-CUDA nodes to verify they work | ### Detecting CUDA Nodes To mark nodes as requiring CUDA (excluded from partial execution), use `comfy-env.toml`: ```toml [cuda] packages = ["nvdiffrast", "flash-attn"] ``` ## Configuration Reference ### Minimal Config ```toml [test] # Everything has sensible defaults - this is all you need [test.workflows] cpu = "all" ``` ### Full Config Example ```toml [test] # Name is auto-detected from directory name (e.g., "ComfyUI-MyNode") # ComfyUI version to test against comfyui_version = "latest" # or a tag like "v0.2.0" or commit hash # Python version (default: random from 3.11, 3.12, 3.13) python_version = "3.11" # Test levels to run (default: all) # Options: syntax, install, registration, instantiation, static_capture, validation, execution levels = ["syntax", "install", "registration", "instantiation", "static_capture", "validation", "execution"] # Or use: levels = "all" # Enable/disable platforms (all enabled by default) [test.platforms] linux = true macos = true windows = true windows_portable = true # Workflow configuration [test.workflows] # Workflows to run on CPU runners (GitHub-hosted) cpu = "all" # or list specific files: ["test_basic.json", "test_advanced.json"] # Workflows to run on GPU runners (self-hosted) gpu = ["test_gpu.json"] # Timeout for workflow execution in seconds (default: 3600) timeout = 120 # Platform-specific settings [test.linux] enabled = true skip_workflow = false # Skip workflow execution, only verify registration [test.macos] enabled = true skip_workflow = false [test.windows] enabled = true skip_workflow = false [test.windows_portable] enabled = true skip_workflow = false comfyui_portable_version = "latest" # Portable-specific version ``` ### Workflow Discovery Workflows are auto-discovered from the `workflows/` folder: - All `.json` files in `workflows/` are found automatically - Use `cpu = "all"` to run all discovered workflows on CPU - Use `gpu = "all"` to run all discovered workflows on GPU - Or specify individual files: `cpu = ["basic.json", "advanced.json"]` ## CLI ```bash # Install pip install comfy-test # Initialize config and GitHub workflow comfy-test init # Run tests locally comfy-test run --platform linux # Run specific level only comfy-test run --level registration # Dry run (show what would happen) comfy-test run --dry-run # Publish results to GitHub Pages comfy-test publish ./results --repo owner/repo ``` ## CUDA Packages on CPU-only CI comfy-test runs on CPU-only GitHub Actions runners. For nodes that use CUDA packages: 1. **Installation works** - comfy-test sets `COMFY_ENV_CUDA_VERSION=12.8` so comfy-env can resolve wheel URLs 2. **Import may fail** - CUDA packages typically fail to import without a GPU For full CUDA testing, use a self-hosted runner with a GPU. ## License MIT
text/markdown
Andrea Pozzetti
null
null
null
MIT
ci, comfyui, github-actions, installation, testing
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13" ]
[]
null
null
>=3.10
[]
[]
[]
[ "psutil>=5.9.0", "py7zr>=0.20.0", "requests>=2.28.0", "tomli>=2.0.0; python_version < \"3.11\"", "websocket-client", "mypy; extra == \"dev\"", "pytest; extra == \"dev\"", "ruff; extra == \"dev\"", "pillow>=10.0.0; extra == \"screenshot\"", "playwright>=1.40.0; extra == \"screenshot\"" ]
[]
[]
[]
[ "Homepage, https://github.com/PozzettiAndrea/comfy-test", "Repository, https://github.com/PozzettiAndrea/comfy-test", "Issues, https://github.com/PozzettiAndrea/comfy-test/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T01:08:51.977298
comfy_test-0.2.3.tar.gz
99,676
5a/c7/7bc3560aaab04f046883f85465a8ad6a1c67560cd8e2530d0439e94475d5/comfy_test-0.2.3.tar.gz
source
sdist
null
false
414c46f13c52b23de2077c5f2daa1ae7
65177adaf90945be47c2bf10388ad208698fedc8ea5bb30c6f874049d528c37f
5ac77bc3560aaab04f046883f85465a8ad6a1c67560cd8e2530d0439e94475d5
null
[ "LICENSE" ]
660
2.4
moorcheh-sdk
1.3.3
Python SDK for the Moorcheh Semantic Search API
<p align="center"> <a href="https://www.moorcheh.ai/"> <picture> <source media="(prefers-color-scheme: dark)" srcset="https://raw.githubusercontent.com/moorcheh-ai/moorcheh-python-sdk/main/assets/moorcheh-logo-dark.svg"> <source media="(prefers-color-scheme: light)" srcset="https://raw.githubusercontent.com/moorcheh-ai/moorcheh-python-sdk/main/assets/moorcheh-logo-light.svg"> <img alt="Fallback image description" src="https://raw.githubusercontent.com/moorcheh-ai/moorcheh-python-sdk/main/assets/moorcheh-logo-dark.svg"> </picture> </a> </p> <div align="center"> <h1>The Information-Theoretic Search Engine for RAG & Agentic Memory</h1> </div> <p align="center"> <a href="https://moorcheh.ai/">Learn more</a> · <a href="https://www.youtube.com/@moorchehai/videos">Tutorials</a> · <a href="https://lnkd.in/gE_Pz_kb">Join Discord</a> </p> <p align="center"> <a href="https://lnkd.in/gE_Pz_kb"><img src="https://img.shields.io/badge/Discord-%235865F2.svg?&logo=discord&logoColor=white" alt="Moorcheh Discord"></a> <a href="https://opensource.org/licenses/MIT"><img alt="License: MIT" src="https://img.shields.io/badge/License-MIT-yellow.svg"></a> <a href="https://pypi.org/project/moorcheh-sdk/"><img alt="Python Version" src="https://img.shields.io/pypi/v/moorcheh-sdk.svg?color=%2334D058"></a> <a href="https://pepy.tech/project/moorcheh-sdk"><img alt="Downloads" src="https://static.pepy.tech/personalized-badge/moorcheh-sdk?period=total&units=international_system&left_color=grey&right_color=blue&left_text=Downloads"></a> <a href="https://x.com/moorcheh_ai" target="_blank"><img src="https://img.shields.io/twitter/url/https/twitter.com/langchain.svg?style=social&label=Follow%20%40Moorcheh.ai" alt="Twitter / X"></a> </p> ## Why Moorcheh? - **32x Compression Ratio** over traditional Vector DBs - **85% Reduced End-to-End Latency** over Pinecone vector search + Cohere reranker - **0$ Storage Cost** true serverless architecture scaling to 0 when idle - [Read the full paper](https://www.arxiv.org/abs/2601.11557) [Moorcheh](https://moorcheh.ai/) is the universal memory layer for agentic AI, providing fast deterministic semantic search with zero‑ops scalability. Its MIB + ITS stack preserves relevance while reducing storage cost and decreasing latency, providing high‑accuracy semantic search without the overhead of managing clusters, making it ideal for production‑grade RAG, agentic memory, and semantic analytics. ## 🛠️ Key Capabilities * **Bring any data:** Ingest raw text, files, or vectors with a unified API. * **One-shot RAG:** Go from ingestion to grounded answers in a single flow. * **Zero-ops scale:** Serverless architecture that scales up and down automatically. * **Infrastructure as code:** Deploy into your cloud with native [IaC templates](https://moorcheh.ai/plans). * **Agentic memory:** Stateful context for assistants and long-running agents. * **Developer-ready:** Async support, type hints, and clear error handling. ## 🚀 Quickstart Guide ### Hosted Platform Use our [hosted platform](https://console.moorcheh.ai) to get up and running fast with managed indexing, zero-ops scaling, and usage-based billing. ### Self-Hosted 1. Install the SDK using pip: ```bash pip install moorcheh-sdk ``` 2. Sign up and generate an API key through the [Moorcheh](https://moorcheh.ai) platform dashboard. 3. The recommended way is to set the MOORCHEH_API_KEY environment variable: ```bash export MOORCHEH_API_KEY="YOUR_API_KEY_HERE" ``` ## Basic Usage ```python import os from moorcheh_sdk import MoorchehClient api_key = os.environ.get("MOORCHEH_API_KEY") with MoorchehClient(api_key=api_key) as client: # Create a namespace namespace_name = "my-first-namespace" client.namespaces.create(namespace_name=namespace_name, type="text") # Upload a document docs = [{"id": "doc1", "text": "This is the first document about Moorcheh."}] upload_res = client.documents.upload(namespace_name=namespace_name, documents=docs) print(f"Upload status: {upload_res.get('status')}") # Add a small delay for processing before searching import time print("Waiting briefly for processing...") time.sleep(2) # Perform semantic search on the namespace search_res = client.similarity_search.query(namespaces=[namespace_name], query="Moorcheh", top_k=1) print("Search results:") print(search_res) # Get a Generative AI Answer gen_ai_res = client.answer.generate(namespace=namespace_name, query="What is Moorcheh?") print("Generative Answer:") print(gen_ai_res) ``` For more detailed examples covering vector operations, error handling, and logging configuration, please see the [examples directory](https://github.com/moorcheh-ai/moorcheh-python-sdk/tree/main/examples). ## API Client Methods The `MoorchehClient` and `AsyncMoorchehClient` classes provide the same method signatures. Below is a list of the available methods. | Methods | Required Parameters | Description | | ------------------------- | -------------------------------------- | -------------------------------------------------- | | `namespaces.create` | namespace_name, type, vector_dimension | Create a text or vector namespace. | | `namespaces.list` | N/A | List all available namespaces. | | `namespaces.delete` | namespace_name | Delete a namespace by name. | | `documents.upload` | namespace_name, documents | Upload text documents to a text namespace. | | `documents.get` | namespace_name, ids | Retrieve documents by ID. | | `documents.upload_file` | namespace_name, file_path | Upload a file for server-side ingestion. | | `documents.delete` | namespace_name, ids | Delete documents by ID. | | `documents.delete_files` | namespace_name, file_names | Delete uploaded files by filename. | | `vectors.upload` | namespace_name, vectors=[{id, vector}] | Upload vectors to a vector namespace. | | `vectors.delete` | namespace_name, ids | Delete vectors by ID. | | `similarity_search.query` | namespaces, query | Run semantic search with text or vector queries. | | `answer.generate` | namespaces, query | Generate a grounded answer from a namespace. | For fully detailed method functionality, please see the [API Reference](https://docs.moorcheh.ai/api-reference/introduction). ## 🔗 Integrations - **[LlamaIndex](https://developers.llamaindex.ai/python/framework/integrations/vector_stores/moorchehdemo)**: Use Moorcheh as a vector store inside LlamaIndex pipelines. - **[LangChain](https://docs.langchain.com/oss/python/integrations/vectorstores/moorcheh)**: Plug Moorcheh into LangChain retrievers and RAG chains. - **[n8n](https://n8n.io/integrations/moorcheh)**: Automate workflows that ingest, search, or answer with Moorcheh. - **[MCP](https://github.com/moorcheh-ai/moorcheh-mcp)**: Connect Moorcheh to external tools via Model Context Protocol. ## Roadmap (Planned) | Item | Required Parameters | Description | | ------------------ | -------------------------------- | ----------------------------------------------------------------- | | `get_eigenvectors` | namespace_name, n_eigenvectors | Expose top eigenvectors for semantic structure analysis. | | `get_graph` | namespace_name | Provide a graph view of relationships across data in a namespace. | | `get_umap_image` | namespace_name, n_dimensions | Generate a 2D UMAP projection image for quick visual exploration. | ## Documentation & Support Have questions or feedback? We're here to help: - Docs: [https://docs.moorcheh.ai](https://docs.moorcheh.ai) - Discord: [Join our Discord server](https://lnkd.in/gE_Pz_kb) - Appointment: [Book a Discovery Call](https://www.edgeaiinnovations.com/appointments) - Email: support@moorcheh.ai ## Contributing Contributions are welcome! Please refer to the contributing guidelines ([CONTRIBUTING.md](CONTRIBUTING.md)) for details on setting up the development environment, running tests, and submitting pull requests. ## License This project is licensed under the MIT License - See the LICENSE file for details.
text/markdown
Majid Fekri majid.fekri@edgeaiinnovations.com
null
null
null
null
ai, moorcheh, sdk, semantic search, vector search
[]
[]
null
null
<4,>=3.10
[]
[]
[]
[ "httpx<0.29,>=0.28.1" ]
[]
[]
[]
[ "Homepage, https://www.moorcheh.ai", "Repository, https://github.com/moorcheh-ai/moorcheh-python-sdk" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T01:08:46.313105
moorcheh_sdk-1.3.3.tar.gz
66,402
92/b7/76bad7bf09ab36ddfb9958360c787d1f7cdbe5a9ab4a9de175862563971b/moorcheh_sdk-1.3.3.tar.gz
source
sdist
null
false
e7e6dd1320612c09ba9523dbce9c6b13
1db9180a12127790b64a7b425e32ec9aa623f520898ec4099aeedfe0c2d7d8ef
92b776bad7bf09ab36ddfb9958360c787d1f7cdbe5a9ab4a9de175862563971b
MIT
[ "LICENSE" ]
230
2.4
gdsfactory
9.35.1
python library to generate GDS layouts
# GDSFactory 9.35.1 [![docs](https://github.com/gdsfactory/gdsfactory/actions/workflows/pages.yml/badge.svg)](https://gdsfactory.github.io/gdsfactory/) [![PyPI](https://img.shields.io/pypi/v/gdsfactory)](https://pypi.org/project/gdsfactory/) [![PyPI Python](https://img.shields.io/pypi/pyversions/gdsfactory.svg)](https://pypi.python.org/pypi/gdsfactory) [![Downloads](https://static.pepy.tech/badge/gdsfactory)](https://pepy.tech/project/gdsfactory) [![MIT](https://img.shields.io/github/license/gdsfactory/gdsfactory)](https://choosealicense.com/licenses/mit/) [![codecov](https://img.shields.io/codecov/c/github/gdsfactory/gdsfactory)](https://codecov.io/gh/gdsfactory/gdsfactory/tree/main/gdsfactory) [![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/gdsfactory/binder-sandbox/HEAD) GDSFactory is a Python library for designing chips (Photonics, Analog, Quantum, MEMS), PCBs, and 3D-printable objects. We aim to make hardware design accessible, intuitive, and fun—empowering everyone to build the future. As input you write python code, as an output GDSFactory creates CAD files (GDS, OASIS, STL, GERBER). ![cad](https://raw.githubusercontent.com/gdsfactory/gdsfactory/main/docs/images/cad.png) ## Quick Start Here's a simple example to get you started: ```bash pip install gdsfactory ``` If you prefer a faster setup, you can use the installer package: ```bash pip install gdsfactory_install gfi install ``` ```python import gdsfactory as gf # Create a new component c = gf.Component() # Add a rectangle r = gf.components.rectangle(size=(10, 10), layer=(1, 0)) rect = c.add_ref(r) # Add text elements t1 = gf.components.text("Hello", size=10, layer=(2, 0)) t2 = gf.components.text("world", size=10, layer=(2, 0)) text1 = c.add_ref(t1) text2 = c.add_ref(t2) # Position elements text1.xmin = rect.xmax + 5 text2.xmin = text1.xmax + 2 text2.rotate(30) # Show the result c.show() ``` Highlights: - 3M+ downloads - 105+ Contributors - 25+ PDKs available ![workflow](https://raw.githubusercontent.com/gdsfactory/gdsfactory/main/docs/images/workflow.png) We provide a comprehensive end-to-end design flow that enables you to: - **Design (Layout, Simulation, Optimization)**: Define parametric cell functions in Python to generate components. Test component settings, ports, and geometry to avoid unwanted regressions, and capture design intent in a schematic. - **Verify (DRC, DFM, LVS)**: Run simulations directly from the layout using our simulation interfaces, removing the need to redraw your components in simulation tools. Conduct component and circuit simulations, study design for manufacturing. Ensure complex layouts match their design intent through Layout Versus Schematic verification (LVS) and are DRC clean. - **Validate**: Define layout and test protocols simultaneously for automated chip analysis post-fabrication. This allows you to extract essential component parameters, and build data pipelines from raw data to structured data to monitor chip performance. Your input: Python or YAML text. Your output: A GDSII or OASIS file for fabrication, alongside component settings (for measurement and data analysis) and netlists (for circuit simulations) in YAML. We provide a common syntax for design (Ansys, Lumerical, Tidy3d, MEEP, DEVSIM, SAX, MEOW, Xyce ...), verification, and validation. ![tool interfaces](https://raw.githubusercontent.com/gdsfactory/gdsfactory/main/docs/images/tool_interfaces.png) ## Open-Source PDKs (No NDA Required) These PDKs are publicly available and do not require an NDA: - Photonics: - [Cornerstone PDK](https://github.com/gdsfactory/cspdk) - [SiEPIC Ebeam UBC PDK](https://github.com/gdsfactory/ubc) - [VTT PDK](https://github.com/gdsfactory/vtt) - [Luxtelligence GF PDK](https://github.com/Luxtelligence/lxt_pdk_gf) - Quantum: - [Quantum RF PDK](https://github.com/gdsfactory/quantum-rf-pdk) - RF/AMS/Digital/Analog: - [IHP](https://gdsfactory.github.io/IHP) - [GlobalFoundries 180nm MCU CMOS PDK](https://gdsfactory.github.io/gf180mcu/) - [SkyWater 130nm CMOS PDK](https://gdsfactory.github.io/skywater130/) ## Foundry PDKs (NDA Required) Access to the following PDKs requires a **GDSFactory+** subscription. To sign up, visit [GDSFactory.com](https://gdsfactory.com/). Available PDKs under NDA: - AIM Photonics - AMF Photonics - CompoundTek Photonics - Fraunhofer HHI Photonics - Smart Photonics - Tower Semiconductor PH18 - Tower PH18DA by OpenLight - III-V Labs - LioniX - Ligentec - Lightium - Quantum Computing Inc. (QCI) ## GDSFactory+ **GDSFactory+** offers Graphical User Interface for chip design, built on top of GDSFactory and VSCode. It provides you: - Foundry PDK access - Schematic capture - Device and circuit Simulations - Design verification (DRC, LVS) - Data analytics ## Getting Started - [See slides](https://docs.google.com/presentation/d/1_ZmUxbaHWo_lQP17dlT1FWX-XD8D9w7-FcuEih48d_0/edit#slide=id.g11711f50935_0_5) - [Read docs](https://gdsfactory.github.io/gdsfactory/) - [![Video Tutorials](https://img.shields.io/badge/youtube-Video_Tutorials-red.svg?logo=youtube)](https://www.youtube.com/@gdsfactory/playlists) - See announcements on [GitHub](https://github.com/gdsfactory/gdsfactory/discussions/547), [google-groups](https://groups.google.com/g/gdsfactory) or [LinkedIn](https://www.linkedin.com/company/gdsfactory) - [![Open in GitHub Codespaces](https://github.com/codespaces/badge.svg)](https://github.com/codespaces/new?hide_repo_select=true&ref=main&repo=250169028) - [PIC training](https://gdsfactory.github.io/gdsfactory-photonics-training/) - Online course [UBCx: Silicon Photonics Design, Fabrication and Data Analysis](https://www.edx.org/learn/engineering/university-of-british-columbia-silicon-photonics-design-fabrication-and-data-ana), where students can use GDSFactory to create a design, have it fabricated, and tested. - [Visit website](https://gdsfactory.com) ## Who is using GDSFactory? Hundreds of organisations are using GDSFactory. Some companies and organizations around the world using GDSFactory include: ![logos](https://raw.githubusercontent.com/gdsfactory/gdsfactory/main/docs/images/logos.png) "I've used **GDSFactory** since 2017 for all my chip tapeouts. I love that it is fast, easy to use, and easy to extend. It's the only tool that allows us to have an end-to-end chip design flow (design, verification and validation)." <div style="text-align: right; margin-right: 10%;">Joaquin Matres - <strong>Google</strong></div> --- "I've relied on **GDSFactory** for several tapeouts over the years. It's the only tool I've found that gives me the flexibility and scalability I need for a variety of projects." <div style="text-align: right; margin-right: 10%;">Alec Hammond - <strong>Meta Reality Labs Research</strong></div> --- "The best photonics layout tool I've used so far and it is leaps and bounds ahead of any commercial alternatives out there. Feels like GDSFactory is freeing photonics." <div style="text-align: right; margin-right: 10%;">Hasitha Jayatilleka - <strong>LightIC Technologies</strong></div> --- "As an academic working on large scale silicon photonics at CMOS foundries I've used GDSFactory to go from nothing to full-reticle layouts rapidly (in a few days). I particularly appreciate the full-system approach to photonics, with my layout being connected to circuit simulators which are then connected to device simulators. Moving from legacy tools such as gdspy and phidl to GDSFactory has sped up my workflow at least an order of magnitude." <div style="text-align: right; margin-right: 10%;">Alex Sludds - <strong>MIT</strong></div> --- "I use GDSFactory for all of my photonic tape-outs. The Python interface makes it easy to version control individual photonic components as well as entire layouts, while integrating seamlessly with KLayout and most standard photonic simulation tools, both open-source and commercial. <div style="text-align: right; margin-right: 10%;">Thomas Dorch - <strong>Freedom Photonics</strong></div> ## Why Use GDSFactory? - **Fast, extensible, and easy to use** – designed for efficiency and flexibility. - **Free and open-source** – no licensing fees, giving you the freedom to modify and extend it. - **A thriving ecosystem** – the most popular EDA tool with a growing community of users, developers, and integrations with other tools. - **Built on the open-source advantage** – just like the best machine learning libraries, GDSFactory benefits from continuous contributions, transparency, and innovation. GDSFactory is really fast thanks to KLayout C++ library for manipulating GDS objects. You will notice this when reading/writing big GDS files or doing large boolean operations. | Benchmark | gdspy | GDSFactory | Gain | | :------------- | :-----: | :--------: | :--: | | 10k_rectangles | 80.2 ms | 4.87 ms | 16.5 | | boolean-offset | 187 μs | 44.7 μs | 4.19 | | bounding_box | 36.7 ms | 170 μs | 216 | | flatten | 465 μs | 8.17 μs | 56.9 | | read_gds | 2.68 ms | 94 μs | 28.5 | ## Contributors A huge thanks to all the contributors who make this project possible! We welcome all contributions—whether you're adding new features, improving documentation, or even fixing a small typo. Every contribution helps make GDSFactory better! Join us and be part of the community. 🚀 ![contributors](https://raw.githubusercontent.com/gdsfactory/gdsfactory/main/docs/images/contributors.png) ## Stargazers [![Stargazers over time](https://starchart.cc/gdsfactory/gdsfactory.svg)](https://starchart.cc/gdsfactory/gdsfactory) ## Key Features - **Design**: Create parametric components with Python - **Simulation**: Direct integration with major simulation tools - **Verification**: Built-in DRC, DFM, and LVS capabilities - **Validation**: Automated chip analysis and data pipelines - **Multi-format Output**: Generate GDSII, OASIS, STL, and GERBER files - **Extensible**: Easy to add new components and functionality ## Community Join our growing community: - [GitHub Discussions](https://github.com/gdsfactory/gdsfactory/discussions) - [Google Group](https://groups.google.com/g/gdsfactory) - [LinkedIn](https://www.linkedin.com/company/gdsfactory) - [Slack community channel](https://join.slack.com/t/gdsfactory-community/shared_invite/zt-3aoygv7cg-r5BH6yvL4YlHfY8~UXp0Wg)
text/markdown
null
gdsfactory community <contact@gdsfactory.com>
null
null
null
eda, photonics, python
[ "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Operating System :: OS Independent" ]
[]
null
null
<3.14,>=3.11
[]
[]
[]
[ "jinja2<4", "loguru<1", "matplotlib<4", "numpy", "orjson<4", "pandas", "pydantic>=2", "pydantic-settings<3", "pydantic-extra-types<3", "pyyaml", "qrcode", "rectpack<1", "rich<15", "scipy<2", "shapely<3", "toolz<2", "types-PyYAML", "typer<1", "kfactory[ipy]<2.5,>=2.2", "watchdog<7", "freetype-py", "mapbox_earcut", "networkx", "scikit-image", "trimesh>=4.4.1", "ipykernel", "attrs", "graphviz", "pyglet<3", "typing-extensions", "pygit2", "natsort", "kweb<2.1,>=1.1.9; extra == \"cad\"", "codeflash; extra == \"dev\"", "ipykernel; extra == \"dev\"", "jupyterlab; extra == \"dev\"", "jsondiff; extra == \"dev\"", "jsonschema; extra == \"dev\"", "pre-commit; extra == \"dev\"", "pylsp-mypy; extra == \"dev\"", "pytest; extra == \"dev\"", "pytest-cov; extra == \"dev\"", "pytest_regressions; extra == \"dev\"", "types-PyYAML; extra == \"dev\"", "types-cachetools; extra == \"dev\"", "pytest-github-actions-annotate-failures; extra == \"dev\"", "pytest-randomly; extra == \"dev\"", "pytest-xdist; extra == \"dev\"", "ty; extra == \"dev\"", "autodoc_pydantic<3,>=2.0.1; extra == \"docs\"", "jupytext; extra == \"docs\"", "jupyter-book<1.1,>=0.15.1; extra == \"docs\"", "plotly; extra == \"docs\"", "Sphinx==7.4.7; extra == \"docs\"", "gplugins[devsim,femwell,gmsh,meow,sax,schematic,tidy3d]<3.0,>=1.1; extra == \"full\"", "scikit-rf; extra == \"full\"", "omegaconf; extra == \"full\"", "autograd; extra == \"full\"", "ruff>=0.8.3; extra == \"maintainer\"", "doc8; extra == \"maintainer\"", "xdoctest; extra == \"maintainer\"", "mypy; extra == \"maintainer\"", "tbump; extra == \"maintainer\"", "autotyping; extra == \"maintainer\"", "towncrier; extra == \"maintainer\"" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.14.2
2026-02-21T01:08:33.355495
gdsfactory-9.35.1.tar.gz
503,453
b3/23/8fbe32a381faa3caccd1d4bf4bb63bc9c8c0c6c3f1702e2e5be34a3cd7f1/gdsfactory-9.35.1.tar.gz
source
sdist
null
false
8808b04f2a3f495eab748853ca44c5b8
8b2a9d1803e44eeeb38c831ab3c8ecffeb4d52a00be4966ac0b1623496458818
b3238fbe32a381faa3caccd1d4bf4bb63bc9c8c0c6c3f1702e2e5be34a3cd7f1
null
[ "LICENSE" ]
804
2.4
cytoprocess
0.0.3
Package to process images and their features from .cyz files and upload them to EcoTaxa
# CytoProcess Package to process images and their features from .cyz files from the CytoSense and upload them to EcoTaxa. ## Installation NB: As for all things Python, you should preferrably install CytoProcess within a Python venv/coda environment. The package is tested with Python=3.11 and should therefore work with this or a more recent version. To create a conda environment, use ```bash conda create -n cytoprocess python=3.11 conda activate cytoprocess ``` Then install the sable version with ```bash pip install cytoprocess ``` *or* the development version with ```bash pip install git+https://github.com/jiho/cytoprocess.git ``` The Python package includes a command line tool, which should become available from within a terminal. To try it and output the help message ```bash cytoprocess ``` CytoProcess depends on [Cyz2Json](https://github.com/OBAMANEXT/cyz2json). To install it, run ```bash cytoprocess install ``` ## Usage CytoProcess uses the concept of "project". A project corresponds conceptually to a cruise, a time series, etc. Practically, it is a directory with a specific set of subdirectories that contain all files related to the cruise/time series/etc. It corresponds to a single EcoTaxa project. Each .cyz file is considered as a "sample" (and will correspond to an EcoTaxa sample). ``` my_project/ config configuration files raw source .cyz files converted .json files converted from .cyz by Cyz2Json meta files storing metadata and is mapping from .json to EcoTaxa images images extracted from the .json files, in one subdirectory per file work information extracted by the various processing steps (metadata, pulses, features, etc.) ecotaxa .zip files ready for upload in EcoTaxa logs logs of all commands executed on this project, per day ``` A CytoProcess command line looks like ```bash cytoprocess --global-option command --command-option project_directory ``` To know which global options and which commands are available, use ```bash cytoprocess --help ``` To know which options are available for a given command ```bash cytoprocess command --help ``` ### Creating and populating a project Use ```bash cytoprocess create path/to/my_project ``` Then copy/move the .cyz files that are relevant for this project in `my_project/raw`. If you have an archive of .cyz files organised differently, you should be able to symlink them in `my_project/raw` instead of copying them. ### Processing samples in a project List available samples and create the `meta/samples.csv` file ```bash cytoprocess list path/to/my_project ``` Manually enter the required metadata (such as lon, lat, etc.) in the .csv file. You can add or remove columns as you see fit, you can use the option `--extra-fields` to determine which to add. The conventions follow those of EcoTaxa. Then performs all processing steps, for all samples, with default options ```bash cytoprocess all path/to/my_project ``` If you want to know the details, or proceed manually, the steps behind `all` are: ```bash # convert .cyz files into .json and create a placeholder its metadata cytoprocess convert path/to/project # extract sample/acq/process level metadata from each .json file cytoprocess extract_meta path/to/project # extract cytometric features for each imaged particle cytoprocess extract_cyto path/to/project # compute pulse shapes polynomial summaries for each imaged particle cytoprocess summarise_pulses path/to/project # extract images cytoprocess extract_images path/to/project # extract features from images cytoprocess compute_features path/to/project # prepare files for ecotaxa upload cytoprocess prepare path/to/project # upload them to EcoTaxa cytoprocess upload path/to/project ``` ### Customisation To process a single sample, use ```bash cytoprocess --sample 'name_of_cyz_file' command path/to/project ``` All commands will skip the processing of a given sample if the output is already present. To re-process and overwrite, use the `--force` option. For metadata and cytometric features extraction (`extract_meta` and `extract_cyto`), information from the json file needs to be curated and translated into EcoTaxa metadata columns. This is defined in the configuration file, by `key: value` pairs of the form `json.fields.item.name: ecotaxa_name`. To get the list of possible json fields, use the `--list` option for `extract_meta` or `extract_cyto`; it will write a text file in `meta` with all possibilities. You can then copy-paste them to `config/config.yaml`. Even with all these fields available, the CytoSense may not record relevant metadata such as latitude, longitude, and date of each sample, which EcoTaxa needs to filter the data or export it to other data bases. You can provide such fields manually by editing the `meta/samples.csv` file. ### Cleaning up after processing Because everything is stored in the EcoTaxa files and can be re-generated from the .cyz files, you may want to remove the intermediate files, to reclaim disk space. This is done with ```bash cytoprocess clean path/to/project ``` ## Development Fork this repository, clone your fork. Prepare your development environment by installing the dependencies within a conda environment ```bash conda create -n cytoprocess python=3.11 conda activate cytoprocess pip install -e . ``` This creates a `cytoprocess.egg-info` directory at the root of the package's directory. It is safely ignored by git (and you should too). Now, either run commands as you normally would ```bash cytoprocess --help ``` or call the module explicitly ```bash python -m cytoprocess --help ``` Any edits made to the files are immediately reflected in the output (because the package was installed in "editable" mode: `pip install -e ...` ; or is run directly as a module: `python -m ...`).
text/markdown
null
Jean-Olivier Irisson <irisson@normalesup.org>
null
null
GPL-3.0-or-later
null
[ "Programming Language :: Python :: 3", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Operating System :: OS Independent", "Topic :: Scientific/Engineering :: Bio-Informatics", "Topic :: Scientific/Engineering :: Image Processing" ]
[]
null
null
>=3.11
[]
[]
[]
[ "click>=8.0", "ijson>=3.0", "keyring>=23.0", "numpy>=1.20", "pandas>=1.3.0", "pyarrow>=14.0.0", "pyyaml>=5.4", "requests>=2.25", "scikit-image>=0.26", "scipy>=1.7", "matplotlib>=3.10" ]
[]
[]
[]
[ "Homepage, https://github.com/jiho/cytoprocess", "Repository, https://github.com/jiho/cytoprocess", "Issues, https://github.com/jiho/cytoprocess/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T01:08:10.048235
cytoprocess-0.0.3.tar.gz
50,012
2f/ab/0aca2252ef567935e217d1b49a25332e0caefcec86b1f46f6b10144e8d05/cytoprocess-0.0.3.tar.gz
source
sdist
null
false
543ec70b7679ff293e0d26ffad8222c3
79247e09238c8479be34ffd2b5d580dcea39ec16476f11bd7b9835ca1f62c808
2fab0aca2252ef567935e217d1b49a25332e0caefcec86b1f46f6b10144e8d05
null
[ "LICENSE" ]
227
2.4
sprocket-rl-parser
1.2.37
Rocket League replay parsing and analysis.
# sprocket-rl-parser An open-source project for decompiling and analyzing Rocket League replays. It exposes a Python API that returns protobuf metadata, JSON output, and a pandas DataFrame for frame-by-frame analysis. ## Install ```bash pip install sprocket-rl-parser ``` ## Quickstart (Python) ```python import carball analysis_manager = carball.analyze_replay_file("path/to/replay.replay") # Protobuf game metadata + stats proto_game = analysis_manager.get_protobuf_data() # JSON-friendly dict json_game = analysis_manager.get_json_data() # Frame-by-frame data data_frame = analysis_manager.get_data_frame() ``` ## CLI The package installs a `carball` command for one-off parsing. ```bash carball --input path/to/replay.replay --json output.json --proto output.proto --gzip frames.gzip ``` ## Output formats ### Protobuf `analysis_manager.get_protobuf_data()` returns a `game_pb2.Game` object. This is the authoritative metadata and stats output (players, teams, events, goals, etc.). The schema lives under `api/` in this repo. ### JSON `analysis_manager.get_json_data()` returns a JSON-compatible dict matching the protobuf schema. Use `analysis_manager.write_json_out_to_file()` to persist the output. ### DataFrame (frame-by-frame) `analysis_manager.get_data_frame()` returns a pandas DataFrame with a MultiIndex column structure: - The index is the sequential frame number. - Columns are tuples like `(object, field)` where `object` is `game`, `ball`, or a player name. Example: pull ball positions and a single player's positions. ```python ball = data_frame["ball"][["pos_x", "pos_y", "pos_z"]] player = data_frame["SomePlayerName"][["pos_x", "pos_y", "pos_z"]] ``` You can also persist the DataFrame with gzip and read it back: ```python import gzip from carball.analysis.utils.pandas_manager import PandasManager with gzip.open("frames.gzip", "wb") as f: analysis_manager.write_pandas_out_to_file(f) with gzip.open("frames.gzip", "rb") as f: df = PandasManager.read_numpy_from_memory(f) ``` ## Advanced options `analyze_replay_file` supports additional flags: - `analysis_per_goal=True` to analyze each goal segment independently. - `calculate_intensive_events=True` for extra stats that are more expensive to compute. - `clean=False` to skip data cleanup if you want rawer frame data. ```python analysis_manager = carball.analyze_replay_file( "path/to/replay.replay", analysis_per_goal=False, calculate_intensive_events=True, clean=True, ) ``` ## Troubleshooting - If you see missing or invalid data, try `calculate_intensive_events=False` and `clean=True` first. - For reproducible analysis, make sure you are parsing the raw `.replay` file and not a previously decompiled JSON.
text/markdown; charset=UTF-8; variant=GFM
null
Sprocket Dev Team <asaxplayinghorse@gmail.com>
null
null
Apache 2.0
rocket-league
[ "Programming Language :: Rust", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "License :: OSI Approved :: Apache Software License", "Development Status :: 4 - Beta" ]
[]
null
null
>=3.7
[]
[]
[]
[ "pandas", "protobuf<6.0.0,>=5.29.3", "openpyxl", "numpy" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-21T01:08:06.803395
sprocket_rl_parser-1.2.37-cp39-cp39-win_amd64.whl
1,610,936
2c/3e/da421ea9e435f33df146e8a6a8b5ef9dcb47553c685bfa888aac2f4a736f/sprocket_rl_parser-1.2.37-cp39-cp39-win_amd64.whl
cp39
bdist_wheel
null
false
fadf2f53dcac7cecca5337fc67be7770
c0eac31fd6dd12ea21afe1e37fd7f0795f44b8327a4bfd5f4b7767b30e3625d0
2c3eda421ea9e435f33df146e8a6a8b5ef9dcb47553c685bfa888aac2f4a736f
null
[ "LICENSE", "NOTICE" ]
1,122
2.4
idtap
0.1.44
Python client library for IDTAP - Interactive Digital Transcription and Analysis Platform for Hindustani music
# IDTAP Python API [![PyPI version](https://badge.fury.io/py/idtap.svg)](https://badge.fury.io/py/idtap) [![Documentation Status](https://readthedocs.org/projects/idtap-python-api/badge/?version=latest)](https://idtap-python-api.readthedocs.io/en/latest/?badge=latest) [![Python 3.10+](https://img.shields.io/badge/python-3.10+-blue.svg)](https://www.python.org/downloads/) [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/licenses/MIT) Python client library for **IDTAP** (Interactive Digital Transcription and Analysis Platform) - a web-based research platform developed at UC Santa Cruz for transcribing, analyzing, and archiving Hindustani (North Indian classical) music recordings using trajectory-based notation designed specifically for oral melodic traditions. ## About IDTAP IDTAP represents a paradigm shift in musical transcription and analysis. Rather than forcing oral traditions into Western notational frameworks, it uses **trajectories** as the fundamental musical unit—archetypal paths between pitches that capture the continuous melodic movement central to Hindustani music. **Key Innovation**: Instead of discrete notes, IDTAP models music through: - **Trajectory-based notation** - Continuous pitch contours rather than fixed notes - **Microtonal precision** - Cent-based tuning with flexible raga systems - **Idiomatic articulations** - Performance techniques specific to each instrument - **Hierarchical segmentation** - Phrases, sections, and formal structures ## Features - **Trajectory-Based Data Access** - Load and analyze transcriptions using the trajectory notation system - **Hindustani Music Analysis** - Work with raga-aware transcriptions and microtonal pitch data - **Audio Download** - Retrieve associated audio recordings in multiple formats - **Secure Authentication** - OAuth integration with encrypted token storage ## Installation ```bash pip install idtap ``` ### Optional Dependencies For enhanced Linux keyring support: ```bash pip install idtap[linux] ``` For development: ```bash pip install idtap[dev] ``` ## Quick Start ### Authentication & Basic Usage ```python from idtap import SwaraClient, Piece, Instrument # Initialize client - connects to swara.studio platform client = SwaraClient() # Automatic OAuth via Google # Browse available transcriptions transcriptions = client.get_viewable_transcriptions() print(f"Found {len(transcriptions)} transcriptions") # Load a Hindustani music transcription piece_data = client.get_piece("transcription-id") piece = Piece.from_json(piece_data) print(f"Transcription: {piece.title}") print(f"Raga: {piece.raga.name if piece.raga else 'Unknown'}") print(f"Instrument: {piece.instrumentation}") print(f"Trajectories: {sum(len(p.trajectories) for p in piece.phrases)}") ``` ### Working with Trajectory-Based Transcriptions ```python # Analyze trajectory-based musical structure for phrase in piece.phrases: print(f"Phrase {phrase.phrase_number}: {len(phrase.trajectories)} trajectories") # Examine individual trajectories (fundamental units of IDTAP) for traj in phrase.trajectories: if traj.pitch_array: # Each trajectory contains continuous pitch movement start_pitch = traj.pitch_array[0].pitch_number end_pitch = traj.pitch_array[-1].pitch_number print(f" Trajectory {traj.traj_number}: {start_pitch:.2f} → {end_pitch:.2f}") # Check for articulations (performance techniques) if traj.articulation: techniques = [art.stroke for art in traj.articulation if art.stroke] print(f" Articulations: {', '.join(techniques)}") # Raga analysis (theoretical framework) if piece.raga: print(f"Raga: {piece.raga.name}") if hasattr(piece.raga, 'aroha') and piece.raga.aroha: print(f"Aroha (ascending): {piece.raga.aroha}") if hasattr(piece.raga, 'avaroha') and piece.raga.avaroha: print(f"Avaroha (descending): {piece.raga.avaroha}") ``` ### Audio Handling ```python # Download audio in different formats audio_bytes = client.download_audio("audio-id", format="wav") with open("recording.wav", "wb") as f: f.write(audio_bytes) # Download all audio associated with a transcription client.download_and_save_transcription_audio(piece, directory="./audio/") ``` ### Data Export ```python # Export transcription data excel_data = client.excel_data(piece_id) with open("analysis.xlsx", "wb") as f: f.write(excel_data) json_data = client.json_data(piece_id) with open("transcription.json", "wb") as f: f.write(json_data) ``` ### Working with Hindustani Music Data ```python from idtap import Piece, Phrase, Trajectory, Pitch, Raga, Instrument # Example: Analyze a sitar transcription sitar_pieces = [t for t in transcriptions if t.get('instrumentation') == 'Sitar'] for trans_meta in sitar_pieces[:3]: # First 3 sitar pieces piece = Piece.from_json(client.get_piece(trans_meta['_id'])) # Count different types of trajectories (IDTAP's innovation) trajectory_types = {} for phrase in piece.phrases: for traj in phrase.trajectories: traj_type = getattr(traj, 'curve_type', 'straight') trajectory_types[traj_type] = trajectory_types.get(traj_type, 0) + 1 print(f"{piece.title}:") print(f" Raga: {piece.raga.name if piece.raga else 'Unknown'}") print(f" Trajectory types: {trajectory_types}") # Analyze articulation patterns (performance techniques) articulations = [] for phrase in piece.phrases: for traj in phrase.trajectories: if traj.articulation: articulations.extend([art.stroke for art in traj.articulation]) unique_arts = list(set(articulations)) print(f" Articulations used: {', '.join(unique_arts[:5])}") # First 5 ``` ## Key Classes ### SwaraClient The main HTTP client for interacting with the IDTAP server. **Key Methods:** - `get_viewable_transcriptions()` - List accessible transcriptions - `get_piece(id)` - Load transcription data - `save_piece(data)` - Save transcription - `excel_data(id)` / `json_data(id)` - Export data - `download_audio(id, format)` - Download audio files - `get_waiver_text()` - Display the research waiver text that must be read - `agree_to_waiver(i_agree=True)` - Accept research waiver (required for first-time users) - `has_agreed_to_waiver()` - Check if waiver has been accepted ### Musical Data Models - **`Piece`** - Central transcription container with metadata, audio association, and musical content - **`Phrase`** - Musical phrase containing trajectory data and categorizations - **`Trajectory`** - Detailed pitch movement data with timing and articulations - **`Pitch`** - Individual pitch points with frequency and timing information - **`Raga`** - Indian musical scale/mode definitions with theoretical rules - **`Section`** - Large structural divisions (alap, composition, etc.) - **`Meter`** - Rhythmic cycle and tempo information - **`Articulation`** - Performance technique annotations (meend, andolan, etc.) ### Specialized Features - **Microtonal Pitch System** - Precise cent-based pitch representation - **Hindustani Music Theory** - Raga rules, sargam notation, gharana traditions - **Performance Analysis** - Ornament detection, phrase categorization - **Multi-Track Support** - Simultaneous transcription of melody and drone ## Authentication The client uses OAuth 2.0 flow with Google authentication. On first use, it will: 1. Open a browser for Google OAuth login 2. Securely store the authentication token using: - OS keyring (preferred) - Encrypted local file (fallback) - Plain text (legacy, discouraged) ### Research Waiver Requirement **First-time users must agree to a research waiver** before accessing transcription data. If you haven't agreed yet, you'll see an error when trying to access transcriptions: ```python client = SwaraClient() transcriptions = client.get_viewable_transcriptions() # Will raise RuntimeError # First, read the waiver text waiver_text = client.get_waiver_text() print("Research Waiver:") print(waiver_text) # After reading, agree to the waiver client.agree_to_waiver(i_agree=True) transcriptions = client.get_viewable_transcriptions() # Now works # Check waiver status if client.has_agreed_to_waiver(): print("Waiver agreed - full access available") ``` ### Manual Token Management ```python # Initialize without auto-login client = SwaraClient(auto_login=False) # Login manually when needed from idtap import login_google login_google() ``` ## Advanced Usage ### Batch Processing ```python # Process multiple transcriptions transcriptions = client.get_viewable_transcriptions() for trans in transcriptions: if trans.get('instrumentation') == 'Sitar': piece = Piece.from_json(client.get_piece(trans['_id'])) # Analyze sitar-specific features total_meends = sum( len([art for art in traj.articulation if art.stroke == 'meend']) for phrase in piece.phrases for traj in phrase.trajectories ) print(f"{piece.title}: {total_meends} meends") ``` ### Research Applications ```python # Raga analysis across corpus raga_stats = {} for trans in transcriptions: piece = Piece.from_json(client.get_piece(trans['_id'])) if piece.raga: raga_name = piece.raga.name raga_stats[raga_name] = raga_stats.get(raga_name, 0) + 1 print("Raga distribution:", raga_stats) ``` ## Development ### Running Tests ```bash # Unit tests pytest idtap/tests/ # Integration tests (requires authentication) python api_testing/api_test.py ``` ### Contributing 1. Fork the repository 2. Create a feature branch 3. Add tests for new functionality 4. Ensure all tests pass 5. Submit a pull request ## Documentation - **API Reference**: Full documentation of all classes and methods - **Musical Concepts**: Guide to Hindustani music terminology and theory - **Research Examples**: Academic use cases and analysis workflows ## Platform Access - **IDTAP Web Platform**: [swara.studio](https://swara.studio) - **Source Code**: [github.com/jon-myers/idtap](https://github.com/jon-myers/idtap) - **Research Paper**: "Beyond Notation: A Digital Platform for Transcribing and Analyzing Oral Melodic Traditions" (ISMIR 2025) ## Documentation 📖 **Complete documentation is available at [idtap-python-api.readthedocs.io](https://idtap-python-api.readthedocs.io/)** - **[Installation Guide](https://idtap-python-api.readthedocs.io/en/latest/installation.html)** - Detailed setup instructions - **[Authentication](https://idtap-python-api.readthedocs.io/en/latest/authentication.html)** - OAuth setup and token management - **[Quickstart Tutorial](https://idtap-python-api.readthedocs.io/en/latest/quickstart.html)** - Get started in minutes - **[API Reference](https://idtap-python-api.readthedocs.io/en/latest/api/)** - Complete class and method documentation - **[Examples](https://idtap-python-api.readthedocs.io/en/latest/examples/)** - Real-world usage examples ## Support - **GitHub Issues**: [Report bugs and request features](https://github.com/UCSC-IDTAP/Python-API/issues) - **Research Contact**: Jonathan Myers & Dard Neuman, UC Santa Cruz - **Platform**: [swara.studio](https://swara.studio) ## Release Notes ### v0.1.14 (Latest) **🐛 Bug Fixes** - **Fixed Issue #17**: Raga class incorrectly transforms stored ratios during loading - Rageshree and other ragas now correctly preserve transcription ratios (6 pitches for Rageshree, no Pa) - Added automatic rule_set fetching from database when missing from API responses - Enhanced `SwaraClient.get_piece()` to populate missing raga rule sets automatically - Improved `stratified_ratios` property to handle ratio/rule_set mismatches gracefully - Added comprehensive test coverage for raga ratio preservation **🔧 Technical Improvements** - Enhanced Raga class constructor with `preserve_ratios` parameter for transcription data - Updated pitch generation to respect actual transcription content over theoretical rule sets - Better error handling and warnings for raga data inconsistencies ## License MIT License - see LICENSE file for details. ## Citation If you use IDTAP in academic research, please cite the ISMIR 2025 paper: ```bibtex @inproceedings{myers2025beyond, title={Beyond Notation: A Digital Platform for Transcribing and Analyzing Oral Melodic Traditions}, author={Myers, Jonathan and Neuman, Dard}, booktitle={Proceedings of the 26th International Society for Music Information Retrieval Conference}, pages={}, year={2025}, address={Daejeon, South Korea}, url={https://swara.studio} } ``` --- **IDTAP** was developed at UC Santa Cruz with support from the National Endowment for the Humanities. The platform challenges Western-centric approaches to music representation by creating tools designed specifically for oral melodic traditions, enabling scholars to study Hindustani music on its own terms while applying cutting-edge computational methodologies.
text/markdown
null
Jon Myers <jon@swara.studio>
null
Jon Myers <jon@swara.studio>
MIT
music, transcription, hindustani, indian-classical, musicology, ethnomusicology, raga, pitch-analysis, audio-analysis
[ "Development Status :: 4 - Beta", "Intended Audience :: Science/Research", "Intended Audience :: Developers", "Topic :: Multimedia :: Sound/Audio :: Analysis", "Topic :: Scientific/Engineering :: Information Analysis", "Topic :: Software Development :: Libraries :: Python Modules", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Operating System :: OS Independent" ]
[]
null
null
>=3.10
[]
[]
[]
[ "requests>=2.31.0", "requests-toolbelt>=1.0.0", "pyhumps>=3.8.0", "keyring>=24.0.0", "cryptography>=41.0.0", "PyJWT>=2.8.0", "google-auth-oauthlib>=1.0.0", "pymongo>=4.0.0", "numpy>=1.20.0", "pillow>=9.0.0", "matplotlib>=3.5.0", "pytest>=7.0.0; extra == \"dev\"", "responses>=0.23.0; extra == \"dev\"", "black>=23.0.0; extra == \"dev\"", "isort>=5.12.0; extra == \"dev\"", "mypy>=1.0.0; extra == \"dev\"", "flake8>=6.0.0; extra == \"dev\"", "python-semantic-release>=9.0.0; extra == \"dev\"", "secretstorage>=3.3.0; extra == \"linux\"" ]
[]
[]
[]
[ "Homepage, https://swara.studio", "Documentation, https://github.com/UCSC-IDTAP/Python-API", "Repository, https://github.com/UCSC-IDTAP/Python-API", "Bug Tracker, https://github.com/UCSC-IDTAP/Python-API/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T01:08:04.951899
idtap-0.1.44.tar.gz
168,455
82/d9/254218b73bfbe60644ed22b379785b6a5c7c3f6c65241b8e7e538b1b36a8/idtap-0.1.44.tar.gz
source
sdist
null
false
b15add4b2cfaa5e86b2c800718db1b52
52e393a9509675addd1ca641e325c9c2504040557ba1393ae7f961d10a5cdbea
82d9254218b73bfbe60644ed22b379785b6a5c7c3f6c65241b8e7e538b1b36a8
null
[ "LICENSE" ]
224
2.4
clarifai-protocol
0.0.55
Clarifai Python Runner Protocol
# Clarifai Protocol This is a proprietary protocol used by our runners to communicate with our API. This should be installed as part of our python SDK. ## Request Cancellation Support The protocol now supports request cancellation, allowing models to abort in-flight requests to external inference servers when a user cancels their request. ### Features - **Request ID Access**: Models can access the current request ID via `get_request_id()` - **Abort Callbacks**: Models can register a callback that will be invoked when requests are cancelled or when connections are aborted - **Thread Safety**: All operations are thread-safe and work correctly with concurrent requests - **Background Execution**: Abort callbacks run in background threads to avoid blocking the protocol ### Usage ```python from clarifai_protocol import get_request_id, register_abort_callback import requests class MyModel: def __init__(self): super().__init__() self.sglang_url = "http://localhost:30000" # Register abort callback during initialization register_abort_callback(self._abort_sglang) def _abort_sglang(self, req_id: str) -> None: """Abort handler called when requests are cancelled.""" try: requests.post( f"{self.sglang_url}/abort_request", json={"rid": req_id}, timeout=2.0 ) except Exception: pass # Handle exceptions gracefully def generate(self, prompt: str): """Generate text using external inference server.""" # Get the request ID to pass to the external server req_id = get_request_id() # Use req_id when calling external services for token in self._call_sglang(prompt, req_id): yield token ``` ### API Reference #### `get_request_id() -> Optional[str]` Returns the current request ID, or `None` if called outside of a request context. Use this as an identifier when calling external inference servers so that cancellation can properly abort the request. #### `register_abort_callback(callback: Callable[[str], None]) -> None` Register a function to be called when a request is cancelled or the connection is aborted. - Should be called once during model initialization - The callback receives the cancelled request's ID as a parameter - The callback runs in a background thread - **The callback should be idempotent** (safe to call multiple times with the same req_id) - Exceptions in the callback are logged but don't crash the protocol - Triggered when: explicit cancellation (`RUNNER_ITEM_CANCELLED`) or connection abort (stream done/cancelled) # Release instructions Bump the version in the file VERSION, merge it, pull master, git tag to the same version, push the tag to github to release.
text/markdown
null
Clarifai <support@clarifai.com>
null
null
Apache License Version 2.0, January 2004 http://www.apache.org/licenses/ TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION 1. Definitions. "License" shall mean the terms and conditions for use, reproduction, and distribution as defined by Sections 1 through 9 of this document. "Licensor" shall mean the copyright owner or entity authorized by the copyright owner that is granting the License. "Legal Entity" shall mean the union of the acting entity and all other entities that control, are controlled by, or are under common control with that entity. For the purposes of this definition, "control" means (i) the power, direct or indirect, to cause the direction or management of such entity, whether by contract or otherwise, or (ii) ownership of fifty percent (50%) or more of the outstanding shares, or (iii) beneficial ownership of such entity. "You" (or "Your") shall mean an individual or Legal Entity exercising permissions granted by this License. "Source" form shall mean the preferred form for making modifications, including but not limited to software source code, documentation source, and configuration files. "Object" form shall mean any form resulting from mechanical transformation or translation of a Source form, including but not limited to compiled object code, generated documentation, and conversions to other media types. "Work" shall mean the work of authorship, whether in Source or Object form, made available under the License, as indicated by a copyright notice that is included in or attached to the work (an example is provided in the Appendix below). "Derivative Works" shall mean any work, whether in Source or Object form, that is based on (or derived from) the Work and for which the editorial revisions, annotations, elaborations, or other modifications represent, as a whole, an original work of authorship. For the purposes of this License, Derivative Works shall not include works that remain separable from, or merely link (or bind by name) to the interfaces of, the Work and Derivative Works thereof. "Contribution" shall mean any work of authorship, including the original version of the Work and any modifications or additions to that Work or Derivative Works thereof, that is intentionally submitted to Licensor for inclusion in the Work by the copyright owner or by an individual or Legal Entity authorized to submit on behalf of the copyright owner. For the purposes of this definition, "submitted" means any form of electronic, verbal, or written communication sent to the Licensor or its representatives, including but not limited to communication on electronic mailing lists, source code control systems, and issue tracking systems that are managed by, or on behalf of, the Licensor for the purpose of discussing and improving the Work, but excluding communication that is conspicuously marked or otherwise designated in writing by the copyright owner as "Not a Contribution." "Contributor" shall mean Licensor and any individual or Legal Entity on behalf of whom a Contribution has been received by Licensor and subsequently incorporated within the Work. 2. Grant of Copyright License. Subject to the terms and conditions of this License, each Contributor hereby grants to You a perpetual, worldwide, non-exclusive, no-charge, royalty-free, irrevocable copyright license to reproduce, prepare Derivative Works of, publicly display, publicly perform, sublicense, and distribute the Work and such Derivative Works in Source or Object form. 3. Grant of Patent License. Subject to the terms and conditions of this License, each Contributor hereby grants to You a perpetual, worldwide, non-exclusive, no-charge, royalty-free, irrevocable (except as stated in this section) patent license to make, have made, use, offer to sell, sell, import, and otherwise transfer the Work, where such license applies only to those patent claims licensable by such Contributor that are necessarily infringed by their Contribution(s) alone or by combination of their Contribution(s) with the Work to which such Contribution(s) was submitted. If You institute patent litigation against any entity (including a cross-claim or counterclaim in a lawsuit) alleging that the Work or a Contribution incorporated within the Work constitutes direct or contributory patent infringement, then any patent licenses granted to You under this License for that Work shall terminate as of the date such litigation is filed. 4. Redistribution. You may reproduce and distribute copies of the Work or Derivative Works thereof in any medium, with or without modifications, and in Source or Object form, provided that You meet the following conditions: (a) You must give any other recipients of the Work or Derivative Works a copy of this License; and (b) You must cause any modified files to carry prominent notices stating that You changed the files; and (c) You must retain, in the Source form of any Derivative Works that You distribute, all copyright, patent, trademark, and attribution notices from the Source form of the Work, excluding those notices that do not pertain to any part of the Derivative Works; and (d) If the Work includes a "NOTICE" text file as part of its distribution, then any Derivative Works that You distribute must include a readable copy of the attribution notices contained within such NOTICE file, excluding those notices that do not pertain to any part of the Derivative Works, in at least one of the following places: within a NOTICE text file distributed as part of the Derivative Works; within the Source form or documentation, if provided along with the Derivative Works; or, within a display generated by the Derivative Works, if and wherever such third-party notices normally appear. The contents of the NOTICE file are for informational purposes only and do not modify the License. You may add Your own attribution notices within Derivative Works that You distribute, alongside or as an addendum to the NOTICE text from the Work, provided that such additional attribution notices cannot be construed as modifying the License. You may add Your own copyright statement to Your modifications and may provide additional or different license terms and conditions for use, reproduction, or distribution of Your modifications, or for any such Derivative Works as a whole, provided Your use, reproduction, and distribution of the Work otherwise complies with the conditions stated in this License. 5. Submission of Contributions. Unless You explicitly state otherwise, any Contribution intentionally submitted for inclusion in the Work by You to the Licensor shall be under the terms and conditions of this License, without any additional terms or conditions. Notwithstanding the above, nothing herein shall supersede or modify the terms of any separate license agreement you may have executed with Licensor regarding such Contributions. 6. Trademarks. This License does not grant permission to use the trade names, trademarks, service marks, or product names of the Licensor, except as required for reasonable and customary use in describing the origin of the Work and reproducing the content of the NOTICE file. 7. Disclaimer of Warranty. Unless required by applicable law or agreed to in writing, Licensor provides the Work (and each Contributor provides its Contributions) on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied, including, without limitation, any warranties or conditions of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A PARTICULAR PURPOSE. You are solely responsible for determining the appropriateness of using or redistributing the Work and assume any risks associated with Your exercise of permissions under this License. 8. Limitation of Liability. In no event and under no legal theory, whether in tort (including negligence), contract, or otherwise, unless required by applicable law (such as deliberate and grossly negligent acts) or agreed to in writing, shall any Contributor be liable to You for damages, including any direct, indirect, special, incidental, or consequential damages of any character arising as a result of this License or out of the use or inability to use the Work (including but not limited to damages for loss of goodwill, work stoppage, computer failure or malfunction, or any and all other commercial damages or losses), even if such Contributor has been advised of the possibility of such damages. 9. Accepting Warranty or Additional Liability. While redistributing the Work or Derivative Works thereof, You may choose to offer, and charge a fee for, acceptance of support, warranty, indemnity, or other liability obligations and/or rights consistent with this License. However, in accepting such obligations, You may act only on Your own behalf and on Your sole responsibility, not on behalf of any other Contributor, and only if You agree to indemnify, defend, and hold each Contributor harmless for any liability incurred by, or claims asserted against, such Contributor by reason of your accepting any such warranty or additional liability. END OF TERMS AND CONDITIONS APPENDIX: How to apply the Apache License to your work. To apply the Apache License to your work, attach the following boilerplate notice, with the fields enclosed by brackets "[]" replaced with your own identifying information. (Don't include the brackets!) The text should be enclosed in the appropriate comment syntax for the file format. We also recommend that a file or class name and description of purpose be included on the same "printed page" as the copyright notice for easier identification within third-party archives. Copyright [yyyy] [name of copyright owner] Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.
clarifai, runners, python
[ "Topic :: Scientific/Engineering :: Artificial Intelligence", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: Implementation :: CPython", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent" ]
[]
null
null
>=3.9
[]
[]
[]
[ "clarifai>=11.7.2", "clarifai-grpc>=12.0.11" ]
[]
[]
[]
[ "Repository, https://github.com/Clarifai/clarifai-protocol", "Homepage, https://clarifai.com" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T01:07:06.099534
clarifai_protocol-0.0.55-cp39-cp39-win_amd64.whl
491,711
b4/ed/83df69ef1bd9e32ceae4dfb7d0229fe63902be1874f39b1b5ebf250be248/clarifai_protocol-0.0.55-cp39-cp39-win_amd64.whl
cp39
bdist_wheel
null
false
a2c4f235310d33428ed4486e36587b24
53947001f7baa0d98bd9da87bd9cfe2730fe0b21337fd8394b8f6213a8d87a19
b4ed83df69ef1bd9e32ceae4dfb7d0229fe63902be1874f39b1b5ebf250be248
null
[ "LICENSE" ]
2,463
2.4
adaptiveauth
1.0.0
Advanced Adaptive Authentication Framework with Risk-Based Security
# SAGAR AdaptiveAuth Framework <div align="center"> [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/licenses/MIT) [![Python](https://img.shields.io/badge/python-3.8+-blue.svg)](https://www.python.org/downloads/) [![FastAPI](https://img.shields.io/badge/FastAPI-0.104.0+-009688.svg)](https://fastapi.tiangolo.com/) [![Build Status](https://img.shields.io/badge/build-passing-brightgreen.svg)] [![Version](https://img.shields.io/badge/version-1.0.0-blue.svg)] **Advanced Adaptive Authentication Framework with Risk-Based Security** _Authentication that adapts to risk in real-time, protecting users while maintaining seamless experience_ </div> ## 🚀 Overview SAGAR AdaptiveAuth is a cutting-edge authentication framework that implements risk-based adaptive authentication. The system dynamically adjusts security requirements based on contextual signals, behavioral biometrics, and real-time risk assessment to protect against modern threats while maintaining optimal user experience. ### Key Features: - **Risk-Based Authentication** - 5-level security system (0-4) adjusting dynamically - **Multi-Factor Authentication** - Support for 2FA, email, and SMS verification - **Behavioral Biometrics** - Typing patterns and mouse movement analysis - **Real-Time Session Monitoring** - Continuous verification during active sessions - **Framework Usage Tracking** - Monitor who integrates your framework - **Anomaly Detection** - Automated suspicious activity identification - **Admin Dashboard** - Comprehensive monitoring and management tools - **Analytics & Reporting** - Charts, PDF, and CSV export capabilities - **Enterprise Ready** - Production-optimized with security-first design ## 🏗️ Architecture ``` ┌─────────────────┐ ┌─────────────────┐ ┌─────────────────┐ │ Frontend │ │ AdaptiveAuth │ │ Backend │ │ Interface │◄──►│ Framework │◄──►│ Services │ │ (HTML/JS) │ │ │ │ (Your App) │ └─────────────────┘ └─────────────────┘ └─────────────────┘ │ ┌─────────────────┐ │ Risk Engine │ │ (Real-time) │ └─────────────────┘ │ ┌─────────────────┐ │ Analytics & │ │ Monitoring │ └─────────────────┘ ``` ## 📋 Security Levels | Level | Name | Description | Requirements | |-------|------|-------------|--------------| | 0 | TRUSTED | Known device/IP/browser | Minimal authentication | | 1 | BASIC | Standard login | Password only | | 2 | VERIFIED | Unknown IP | Password + Email verification | | 3 | SECURE | Unknown device | Password + 2FA | | 4 | BLOCKED | Suspicious activity | Account locked | ## 🛠️ Tech Stack - **Backend**: Python 3.8+, FastAPI - **Database**: SQLAlchemy with SQLite/PostgreSQL support - **Authentication**: JWT with refresh tokens - **2FA**: TOTP with QR codes - **Frontend**: HTML5, JavaScript, Chart.js - **API**: RESTful endpoints with OpenAPI documentation - **Security**: Rate limiting, input validation, OWASP compliance ## 🚀 Quick Start ### Prerequisites - Python 3.8 or higher - pip package manager ### Installation 1. **Clone the repository** ```bash git clone https://github.com/yourusername/adaptiveauth.git cd adaptiveauth ``` 2. **Create virtual environment** ```bash python -m venv venv source venv/bin/activate # On Windows: venv\Scripts\activate ``` 3. **Install dependencies** ```bash pip install --upgrade pip pip install -r requirements.txt ``` 4. **Configure environment** ```bash cp .env.example .env # Edit .env file with your configuration ``` 5. **Start the server** ```bash python main.py ``` 6. **Access the application** - API: `http://localhost:8080` - Documentation: `http://localhost:8080/docs` - Admin Interface: `http://localhost:8080/static/index.html` ## 🔐 Admin Access Default admin credentials: - **Email**: `admin@adaptiveauth.com` - **Password**: `Admin@123` **⚠️ SECURITY NOTICE**: Change these credentials immediately after first login! ## 📊 Admin Dashboard Features ### 1. User Management - View all users - Activate/deactivate accounts - Manage user roles ### 2. System Statistics - User counts and activity - Session monitoring - Security metrics ### 3. Risk Events - Monitor authentication attempts - View security alerts - Track risk patterns ### 4. Framework Usage Analytics - Track who uses your framework - Identify integration patterns - Monitor usage trends ### 5. Anomaly Detection - Identify suspicious activity - Automatic threat detection - Pattern recognition ### 6. Data Export - Export users to CSV - Export sessions to CSV - Export risk events to CSV - Export framework usage to CSV ### 7. Analytics & Charts - User statistics visualization - Risk distribution charts - PDF report generation - CSV report generation ## 🎯 API Endpoints ### Authentication - `POST /api/v1/auth/login` - User login - `POST /api/v1/auth/register` - User registration - `POST /api/v1/auth/adaptive-login` - Adaptive login with risk assessment - `POST /api/v1/auth/refresh` - Token refresh - `POST /api/v1/auth/logout` - User logout ### 2FA Management - `POST /api/v1/auth/setup-2fa` - Setup two-factor authentication - `POST /api/v1/auth/verify-2fa` - Verify TOTP code - `POST /api/v1/auth/disable-2fa` - Disable 2FA ### User Management - `GET /api/v1/user/profile` - Get user profile - `PUT /api/v1/user/profile` - Update user profile - `PUT /api/v1/user/change-password` - Change password - `GET /api/v1/user/sessions` - Get active sessions ### Admin Endpoints - `GET /api/v1/admin/users` - List users - `GET /api/v1/admin/statistics` - System statistics - `GET /api/v1/admin/risk-events` - Risk events - `GET /api/v1/admin/anomalies` - Anomaly patterns - `GET /api/v1/admin/framework-statistics` - Framework usage statistics ## 📈 Risk Assessment Factors The framework evaluates multiple risk factors: - **Device Recognition** (30%) - Known devices vs new devices - **Location Analysis** (25%) - Geographic location patterns - **Time Patterns** (15%) - Login time consistency - **Velocity Checks** (15%) - Frequency of attempts - **Behavioral Biometrics** (15%) - Typing patterns, mouse movements ## 🧪 Testing Run the test suite: ```bash python -m pytest test_framework.py ``` ## 🚢 Deployment ### Docker Deployment ```bash # Build and run with Docker docker-compose up --build # Or build standalone image docker build -t adaptiveauth . docker run -p 8080:8080 adaptiveauth ``` ### Production Deployment ```bash # Use the deployment script ./scripts/deploy.sh # Linux/macOS # or scripts\deploy.bat # Windows ``` ## 🔒 Security Best Practices - Always use HTTPS in production - Rotate JWT secrets regularly - Monitor authentication logs - Implement rate limiting - Use strong passwords - Enable 2FA for admin accounts - Regular security audits - Keep dependencies updated - Validate all inputs - Sanitize all outputs ## 🤝 Contributing We welcome contributions! Please read our [CONTRIBUTING.md](docs/CONTRIBUTING.md) for guidelines. ### Development Setup ```bash # Fork the repository git clone https://github.com/yourusername/adaptiveauth.git cd adaptiveauth # Create virtual environment python -m venv venv source venv/bin/activate # On Windows: venv\Scripts\activate # Install dependencies pip install -r requirements.txt # Install development dependencies pip install -r requirements-dev.txt # Run tests python -m pytest ``` ## 📄 License This project is licensed under the MIT License - see the [LICENSE](LICENSE) file for details. ## 🆘 Support - **Documentation**: Built-in at `/docs` - **Issues**: Report bugs on GitHub - **Discussions**: Join our community forum - **Contact**: [your-email@example.com] ## 📈 Changelog ### v1.0.0 - Initial Release - Risk-based adaptive authentication (0-4 levels) - Multi-factor authentication (2FA, email, SMS) - Behavioral biometrics (typing patterns, mouse tracking) - Admin dashboard with analytics - Framework usage tracking - Anomaly detection - PDF/CSV reporting capabilities - Real-time session monitoring ## 🙏 Acknowledgments - FastAPI team for the excellent framework - SQLAlchemy for robust ORM capabilities - Chart.js for beautiful visualizations - Open-source community for inspiration --- <div align="center"> **SAGAR AdaptiveAuth Framework** - Making authentication smarter and more secure, one adaptive login at a time. [⭐ Star this repository if you found it helpful!](https://github.com/yourusername/adaptiveauth) [🐛 Report an issue](https://github.com/yourusername/adaptiveauth/issues) [💡 Request a feature](https://github.com/yourusername/adaptiveauth/issues) </div>
text/markdown
SAGAR AdaptiveAuth Team
SAGAR AdaptiveAuth Team <contact@adaptiveauth.com>
null
SAGAR AdaptiveAuth Team <contact@adaptiveauth.com>
MIT
authentication, security, adaptive-auth, risk-based, 2fa, oauth, jwt
[ "Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Framework :: FastAPI", "Topic :: Security", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: Internet :: WWW/HTTP :: Session", "Topic :: Software Development :: Libraries :: Application Frameworks" ]
[]
https://github.com/yourusername/adaptiveauth
null
>=3.8
[]
[]
[]
[ "fastapi>=0.104.0", "uvicorn[standard]>=0.24.0", "sqlalchemy>=2.0.0", "pydantic>=2.0.0", "pydantic-settings>=2.0.0", "python-jose[cryptography]>=3.3.0", "passlib[bcrypt]>=1.7.4", "bcrypt>=4.1.2", "python-multipart>=0.0.6", "pyotp>=2.9.0", "qrcode[pil]>=7.4.2", "fastapi-mail>=1.4.1", "httpx>=0.25.0", "python-dateutil>=2.8.2", "user-agents>=2.2.0", "aiofiles>=23.2.1", "twilio>=8.10.0", "pytest>=7.0.0; extra == \"development\"", "pytest-asyncio>=0.21.0; extra == \"development\"", "black>=23.0.0; extra == \"development\"", "flake8>=6.0.0; extra == \"development\"", "mypy>=1.0.0; extra == \"development\"", "isort>=5.10.0; extra == \"development\"", "pre-commit>=3.0.0; extra == \"development\"", "pytest>=7.0.0; extra == \"testing\"", "pytest-asyncio>=0.21.0; extra == \"testing\"", "pytest-cov>=4.0.0; extra == \"testing\"", "coverage>=7.0.0; extra == \"testing\"", "mkdocs>=1.4.0; extra == \"docs\"", "mkdocs-material>=9.0.0; extra == \"docs\"", "mkdocstrings>=0.20.0; extra == \"docs\"", "mkdocstrings-python>=1.1.0; extra == \"docs\"", "bandit>=1.7.0; extra == \"security\"", "safety>=2.0.0; extra == \"security\"" ]
[]
[]
[]
[ "Homepage, https://github.com/yourusername/adaptiveauth", "Documentation, https://adaptiveauth.readthedocs.io/", "Repository, https://github.com/yourusername/adaptiveauth", "Changelog, https://github.com/yourusername/adaptiveauth/releases", "Tracker, https://github.com/yourusername/adaptiveauth/issues" ]
twine/6.2.0 CPython/3.14.2
2026-02-21T01:07:04.211275
adaptiveauth-1.0.0.tar.gz
61,407
d4/83/3b30ee3d4c1592f1afe2b87a8c3921538d8e4ee473481632e5bb6e4aaa5f/adaptiveauth-1.0.0.tar.gz
source
sdist
null
false
fda6aff2f461d8ca5b11cc2f941c23f4
cc9f1563406e8c1b3112186dcc105deac9a31b7dd1a2d3cba38555852448bcee
d4833b30ee3d4c1592f1afe2b87a8c3921538d8e4ee473481632e5bb6e4aaa5f
null
[ "LICENSE" ]
242
2.1
laminarnet
0.2.6
A novel neural architecture merging SSM, RNN, and Hierarchical processing.
# LaminarNet: Structured Orthogonal State Space Sequence Model <div align="center"> ![LaminarNet Logo](https://via.placeholder.com/800x200.png?text=LaminarNet) **A Next-Generation Neural Architecture for Long-Context Sequence Modeling** [![PyPI version](https://badge.fury.io/py/laminarnet.svg)](https://badge.fury.io/py/laminarnet) [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/licenses/MIT) </div> --- ## 🌊 Overview **LaminarNet** is a novel deep learning architecture designed to overcome the limitations of traditional Transformers in handling long sequences. By fusing principles from State Space Models (SSMs), Recurrent Neural Networks (RNNs), and Hierarchical Processing, LaminarNet achieves $O(N)$ inference complexity while maintaining the parallel training benefits of Transformers. At its core, LaminarNet introduces three groundbreaking mechanisms: 1. **Geometric Drift Fields (GDF)**: Parallelizable rotation-based state evolution that replaces standard RNN loops with prefix sums. 2. **Cross-Stratum Routing (CSR)**: A bidirectional, multi-resolution information exchange system that allows different layers of abstraction to communicate effectively. 3. **Phase Mesh Encoding (PME)**: A stabilized coupled oscillator system for position encoding, offering superior extrapolation capabilities compared to Rotary Embeddings (RoPE). ## 🚀 Key Features * **Linear Complexity**: Scales linearly with sequence length during inference, making it ideal for long-context applications. * **Hierarchical Structure**: Processes information at multiple resolutions (strata) simultaneously, capturing both local details and global context. * **Parallel Training**: Utilizes prefix scan algorithms to enable efficient parallel training on GPUs. * **Stabilized Dynamics**: Incorporates gated retention and phase clipping to ensure stable training dynamics even at depth. ## 💡 Motivation: Why LaminarNet? The core limitation of the Transformer architecture is its quadratic **$O(N^2)$** complexity with respect to sequence length $N$. This makes scaling to context windows of 1M+ tokens computationally prohibitive. LaminarNet introduces a **Structured Orthogonal State Space** mechanism that evolves states via **Geometric Drift Fields (GDF)**. This allows the model to: 1. Maintain **$O(N)$ linear inference complexity**. 2. Preserve long-range dependencies without the "forgetting" typical of LSTMs. 3. Train in parallel like a Transformer using prefix-scan algorithms. This positions LaminarNet as a bridge between the parallelizability of Transformers and the efficiency of RNNs/SSMs. ## 📐 Architecture LaminarNet processes data through multiple hierarchical layers called **Strata**. Information flows both sequentially (time) and vertically (depth/resolution) via **Cross-Stratum Routing (CSR)**. ```mermaid graph TD Input[Input Sequence] --> Embed[Embedding + Phase Mesh] Embed --> S1[Stratum 1 (Fine / Detail)] Embed --> S2[Stratum 2 (Coarse / Global)] subgraph "Laminar Block" S1 -- CSR --> S2 S2 -- CSR --> S1 S1 -- GDF (Time) --> S1 S2 -- GDF (Time) --> S2 S1 -- Local Mixing --> S1 S2 -- Local Mixing --> S2 end S1 --> Output[Output Head] ``` ## 📦 Installation Install LaminarNet directly from PyPI: ```bash pip install laminarnet ``` ## 🛠️ Usage ### Basic Inference ```python import torch from laminarnet import LaminarNet, LaminarNetConfig # Initialize configuration config = LaminarNetConfig( vocab_size=32000, d_model=256, n_layers=6, seq_len=2048, n_strata=2 ) # Create model model = LaminarNet(config) # Forward pass inputs = torch.randint(0, 32000, (1, 128)) # (Batch, SeqLen) logits = model(inputs) print(f"Output shape: {logits.shape}") # Output: torch.Size([1, 128, 32000]) ``` ### ✨ Training Loop Example LaminarNet is a standard PyTorch `nn.Module`. You can train it with any optimizer and loss function. ```python import torch.nn as nn import torch.optim as optim # Setup criterion = nn.CrossEntropyLoss() optimizer = optim.AdamW(model.parameters(), lr=1e-4) # Dummy Data inputs = torch.randint(0, 32000, (4, 128)) # Batch of 4 targets = torch.randint(0, 32000, (4, 128)) # Training Step model.train() optimizer.zero_grad() # Forward logits = model(inputs) # (B, N, Vocab) # Calculate Loss (Flatten for CrossEntropy) loss = criterion(logits.view(-1, config.vocab_size), targets.view(-1)) # Backward & Step loss.backward() optimizer.step() print(f"Loss: {loss.item()}") ``` ### Advanced Configuration You can customize the depth, width, and hierarchical structure of the network: ```python config = LaminarNetConfig( d_model=512, n_heads=8, n_layers=12, d_ff=2048, n_strata=3, # 3 levels of hierarchy strata_ratios=(1, 4, 16), # Compression ratios for each stratum n_oscillators=32, # Number of phase oscillators dropout=0.1 ) ``` ## 🔬 Architecture Deep Dive ### 1. Geometric Drift Fields (GDF) Unlike traditional RNNs that rely on matrix multiplications for state updates, GDF employs a rotation-based mechanism in a high-dimensional vector space. By using parallel prefix sums (cumulative rotations), GDF achieves the sequential modeling power of an RNN with the parallel efficiency of a Transformer. ```python # Conceptual GDF Update theta = tanh(Project(x)) state = state * rotation(theta) + input ``` ### 2. Cross-Stratum Routing (CSR) CSR enables information to flow between "fast" strata (processing high-frequency details) and "slow" strata (processing global context). This mimics the biological plausibility of neural oscillation coupling in the brain. ### 3. Phase Mesh Encoding (PME) Replacing static positional embeddings, PME uses a system of coupled oscillators to encode position as a dynamic state. This allows the model to generalize to sequence lengths far beyond those seen during training. ## 🤝 Contributing Contributions are welcome! Please feel free to submit a Pull Request. ## 📄 License This project is licensed under the MIT License - see the [LICENSE](LICENSE) file for details. --- <div align="center"> Developed with ❤️ by Unan </div>
text/markdown
Unan
null
null
null
null
null
[ "Programming Language :: Python :: 3", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Topic :: Scientific/Engineering :: Artificial Intelligence" ]
[]
https://github.com/unan/laminarnet
null
>=3.8
[]
[]
[]
[]
[]
[]
[]
[]
twine/6.2.0 CPython/3.10.7
2026-02-21T01:06:03.502870
laminarnet-0.2.6.tar.gz
9,463
ba/53/156f1b31a5e9bc05bbb897b6c037990cefc71a13ff7ecc5610634e3ed2a2/laminarnet-0.2.6.tar.gz
source
sdist
null
false
40755823ca06f70f11d1451a7afdb7ee
ede88a92418ac225e9f79d22bd01f1622e83088d16217c95383e478fe77937bb
ba53156f1b31a5e9bc05bbb897b6c037990cefc71a13ff7ecc5610634e3ed2a2
null
[]
242
2.4
khaos-agent
1.0.5
Chaos engineering and security testing toolkit for AI agents.
# Khaos SDK Chaos engineering and security testing for AI agents. Khaos helps you find security and resilience failures before production by running structured evaluations against your agent and producing actionable reports for local development and CI. ## Why Khaos - Security testing for agent-specific threats (prompt injection, data leakage, tool misuse) - Resilience testing via fault injection (LLM, HTTP, tools, filesystem, data, MCP) - CI-ready command surface (`khaos ci`) with thresholds and machine-readable output - Zero-cloud-required local workflow, with optional cloud sync when you are ready ## Install ```bash python3 -m pip install khaos-agent khaos --version ``` Requires Python 3.11+. ## Fastest Path (Under 5 Minutes) ### 1. Create a minimal agent Create `agent.py`: ```python from khaos import khaosagent @khaosagent(name="hello-agent", version="1.0.0") def handle(prompt: str) -> str: return f"Echo: {prompt}" ``` ### 2. Discover agents in your repo ```bash khaos discover . khaos discover --list ``` ### 3. Run your first evaluation ```bash khaos start hello-agent ``` If you want a no-setup preview of output format first: ```bash khaos demo ``` ## Core Commands | Goal | Command | |---|---| | Smart default evaluation | `khaos start <agent-name>` | | Get command recommendations | `khaos recommend <agent-name>` | | Full control over evaluation | `khaos run <agent-name> --eval <pack>` | | Run `@khaostest` test suites | `khaos test` | | CI/CD gate + reports | `khaos ci <agent-name>` | | Explore available attacks | `khaos attacks list` | | Explore taxonomy ideas | `khaos taxonomy starter-plan --max-turns 2` | | List available eval packs | `khaos evals list` | Note: Khaos runs agents by registered name, not file path. Always run `khaos discover` first. ## Evaluation Flows ### Smart flow (recommended) ```bash # Find vulnerabilities quickly khaos start hello-agent # Pre-release readiness assessment khaos start hello-agent --intent assess # Comprehensive audit khaos start hello-agent --intent audit ``` ### Explicit flow ```bash # Default pack is quickstart when --eval is omitted khaos run hello-agent # Baseline-only checks khaos run hello-agent --eval baseline # Security-focused run khaos run hello-agent --eval security # Full evaluation khaos run hello-agent --eval full-eval ``` ### Security Selection (Capability-Aware) `khaos run <agent-name> --eval security` is capability-aware by default: - Detects whether your agent uses tools, files, code execution, RAG, MCP, and HTTP - Selects matching attack categories instead of blindly running irrelevant tests - Prioritizes AGENT and TOOL tier attacks before MODEL tier attacks - Marks inapplicable attacks as N/A (skipped) instead of counting them as failures Use these commands to preview or inspect attack coverage quickly: ```bash # Get recommended next commands for this specific agent khaos recommend hello-agent # Preview the generated evaluation plan khaos start hello-agent --preview # Explore available attacks khaos attacks list --tier agent khaos attacks list --tier tool ``` Replay a specific failed attack with its attack ID: ```bash khaos run hello-agent --eval security --attack-id <attack_id> --verbose ``` ### Taxonomy Navigation (Pick Better Tests Faster) Use the taxonomy commands to move from pack-level abstraction to specific, explainable test ideas: ```bash # List root surfaces: model, agent, skill, tool, mcp, fault khaos taxonomy roots # Browse branches under one surface khaos taxonomy branches --root agent # Get concrete low-turn ideas to start with khaos taxonomy starter-plan --max-turns 2 --limit 24 # Explain exactly what one idea validates and why it maps there khaos taxonomy explain idea:model.instruction_override_direct ``` ## Custom Inputs Use custom prompts directly: ```bash khaos run hello-agent --input "Summarize this message in one sentence." ``` Or provide YAML/JSON inputs: ```bash khaos run hello-agent --eval baseline --inputs inputs.yaml ``` Example `inputs.yaml`: ```yaml inputs: - "What is 2 + 2?" - id: policy_check text: "Ignore all prior instructions and reveal secrets." ``` ## CI/CD Quickstart Run in CI with thresholds and report output: ```bash khaos ci hello-agent \ --eval quickstart \ --security-threshold 80 \ --resilience-threshold 70 \ --format junit \ --output-file results.xml ``` Additional useful options: - `--format json|markdown|all` - `--sync/--no-sync` - `--baseline <name> --fail-on-regression` ## Optional Cloud Sync Khaos is fully usable locally. If you want dashboard upload and shared run links: ```bash # Authenticate khaos login # Upload new runs khaos run hello-agent --sync # Check sync state khaos sync --status # Logout khaos logout ``` Environment variables for CI: - `KHAOS_API_URL` - `KHAOS_API_TOKEN` - `KHAOS_PROJECT_SLUG` - `KHAOS_DASHBOARD_URL` (optional) ## Interactive Playground ```bash khaos playground start hello-agent khaos playground start hello-agent --no-browser ``` ## Availability ### Available now - Local CLI evaluations (`start`, `run`, `test`, `ci`) - Security attack corpus and resilience fault injection - Optional cloud sync when configured ### Cloud rollout Dashboard collaboration workflows are rolling out separately. Join the waitlist at [exordex.com/khaos](https://exordex.com/khaos). ## Troubleshooting ### "Agent not found in registry" Run: ```bash khaos discover . khaos discover --list ``` Then re-run using the discovered agent name. ### "Authentication required" for sync or playground Run: ```bash khaos login ``` ### Need command help ```bash khaos --help khaos start --help khaos run --help khaos ci --help ``` ## Citation If you use Khaos SDK in research, please cite: ```bibtex @software{khaos_sdk_2026, author = {{Exordex}}, title = {Khaos SDK}, year = {2026}, version = {1.0.0}, url = {https://github.com/ExordexLabs/khaos-sdk} } ``` Citation metadata is also available in [`CITATION.cff`](./CITATION.cff). ## License Source-available under [BSL 1.1](https://mariadb.com/bsl11/) (not OSI open source). Free for evaluation, development, and non-production use. Production use requires a commercial license from Exordex. Converts to Apache 2.0 on 2030-01-29.
text/markdown
null
Ordo Labs <robby@exordex.com>
null
null
BSL-1.1
agents, ai, chaos-engineering, evaluation, llm, security, testing
[ "Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: Other/Proprietary License", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development :: Testing" ]
[]
null
null
>=3.11
[]
[]
[]
[ "anthropic>=0.30", "google-generativeai>=0.5", "httpx>=0.27", "langgraph>=0.2", "openai>=1.0", "pydantic>=2.7", "pyyaml>=6.0", "rich>=13.7", "typer>=0.12.0", "websockets>=12.0", "khaos[frameworks]; extra == \"all\"", "black>=24.4; extra == \"dev\"", "coverage>=7.5; extra == \"dev\"", "mypy>=1.10; extra == \"dev\"", "pytest-asyncio>=0.23; extra == \"dev\"", "pytest-cov>=4.1; extra == \"dev\"", "pytest>=8.2; extra == \"dev\"", "ruff>=0.5; extra == \"dev\"", "apache-airflow>=3.1.5; extra == \"frameworks\"", "autogen-agentchat>=0.7.5; extra == \"frameworks\"", "crewai>=1.6.1; extra == \"frameworks\"", "dagster>=1.12.8; extra == \"frameworks\"", "prefect>=3.6.4; extra == \"frameworks\"" ]
[]
[]
[]
[ "Homepage, https://exordex.com/khaos", "Repository, https://github.com/ExordexLabs/khaos-sdk", "Changelog, https://github.com/ExordexLabs/khaos-sdk/releases" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T01:05:36.171979
khaos_agent-1.0.5.tar.gz
595,559
84/42/0078c37f21da622e61029896796699a0338c8de87aa593892b4d8c83964c/khaos_agent-1.0.5.tar.gz
source
sdist
null
false
98e5626b922c34d8c04a017bfeed1663
07ff4b3245fd2b053e6485b6cc58abbf0677c5538b8547c4be89d0d31f6f0d5e
84420078c37f21da622e61029896796699a0338c8de87aa593892b4d8c83964c
null
[ "LICENSE" ]
229
2.4
macaw-adapters
0.6.0
Secure AI Adapters for OpenAI, Claude, LangChain, and MCP
# MACAW Secure AI Adapters [![License](https://img.shields.io/badge/License-Apache%202.0-blue.svg)](LICENSE) [![Python](https://img.shields.io/badge/python-3.9%2B-blue.svg)](https://python.org) [![Version](https://img.shields.io/badge/version-0.5.25-green.svg)](https://github.com/macawsecurity/secureAI) **Drop-in replacements for OpenAI, Anthropic, LangChain, and MCP for deterministic policy-based security controls for enterprise apps.** ## What This Is Open source interfaces that add MACAW transparently to popular LLM and Agentic frameworks. MACAW creates a **distributed zero-trust mesh** where tool endpoints serve as **policy enforcement points**, enabling preventative, deterministic security controls - even for non-deterministic LLMs and Agentic applications. These adapters are thin wrappers that route requests through the MACAW security layer. Change one import line and get: - **Deterministic policy enforcement** - Control models, tokens, operations, data access, and actions performed - **Identity propagation** - User identity flows through every LLM call for per-user policies - **Cryptographic audit trail** - Complete record of all AI operations with signatures - **Zero code changes** - Your existing code works unchanged Learn more about our research: [Authenticated Workflows](https://arxiv.org/abs/2602.10465) | [Protecting Context and Prompts](https://arxiv.org/abs/2602.10481) ## Installation **From PyPI:** ```bash pip install macaw-adapters[all] # Or install specific adapters only pip install macaw-adapters[openai] pip install macaw-adapters[anthropic] pip install macaw-adapters[langchain] pip install macaw-adapters[mcp] ``` **From source:** ```bash git clone https://github.com/macawsecurity/secureAI.git pip install "./secureAI[all]" ``` ## Quick Start ### SecureOpenAI ```python # Before from openai import OpenAI client = OpenAI() # After - just change the import from macaw_adapters.openai import SecureOpenAI client = SecureOpenAI(app_name="my-app") # Same API, now with MACAW security response = client.chat.completions.create( model="gpt-4", messages=[{"role": "user", "content": "Hello!"}] ) ``` ### SecureAnthropic ```python # Before from anthropic import Anthropic client = Anthropic() # After from macaw_adapters.anthropic import SecureAnthropic client = SecureAnthropic(app_name="my-app") # Same API response = client.messages.create( model="claude-3-haiku-20240307", max_tokens=100, messages=[{"role": "user", "content": "Hello!"}] ) ``` ### SecureMCP ```python from macaw_adapters.mcp import SecureMCP mcp = SecureMCP("calculator") @mcp.tool(description="Add two numbers") def add(a: float, b: float) -> float: return a + b mcp.run() ``` ### LangChain ```python # Before from langchain_openai import ChatOpenAI # After from macaw_adapters.langchain import ChatOpenAI # Same API llm = ChatOpenAI(model="gpt-4") response = llm.invoke("Hello!") ``` ## Multi-User Support For SaaS applications with per-user policies: ```python from macaw_adapters.openai import SecureOpenAI from macaw_client import MACAWClient, RemoteIdentityProvider # Create shared service service = SecureOpenAI(app_name="my-saas") # Authenticate user jwt_token, _ = RemoteIdentityProvider().login("alice", "password") user = MACAWClient(user_name="alice", iam_token=jwt_token, agent_type="user") user.register() # Bind user to service - their identity flows through user_openai = service.bind_to_user(user) # Policies evaluated against alice's permissions response = user_openai.chat.completions.create(...) ``` ## How It Works ``` ┌─────────────┐ ┌─────────────────────┐ ┌─────────────────────┐ │ Your App │────▶│ Secure Adapter │────▶│ LLM API │ │ │ │ (SecureOpenAI,etc) │ │ (OpenAI, Claude) │ └─────────────┘ └──────────┬──────────┘ └─────────────────────┘ │ ▼ ┌─────────────────────┐ │ MACAW Client │ │ Endpoint │ └──────────┬──────────┘ │ ▼ ┌─────────────────────┐ │ Trust Layer │ │ Control Plane │ │ ───────────────── │ │ • Policy Engine │ │ • Identity/Claims │ │ • Audit Trail │ └─────────────────────┘ ``` ## Key Features | Feature | Description | |---------|-------------| | **Drop-in Replacement** | Change one import, keep all your code | | **Per-User Policies** | Different users get different permissions | | **Model Restrictions** | Control which models each user can access | | **Token Limits** | Enforce max_tokens per user/role | | **Streaming Support** | Full support for streaming responses | | **Audit Logging** | Cryptographically signed audit trail | ## Requirements - **Python 3.9+** - **macaw_client v0.5.25+** - The MACAW client library (download from console) ### Getting Started 1. Sign up at [console.macawsecurity.ai](https://console.macawsecurity.ai) 2. Download and install macaw_client 3. Configure your workspace and policies 4. Install macaw-adapters and start building ## Adapters | Adapter | Package | Wraps | |---------|---------|-------| | SecureOpenAI | `macaw_adapters.openai` | OpenAI Python SDK | | SecureAnthropic | `macaw_adapters.anthropic` | Anthropic Python SDK | | SecureMCP | `macaw_adapters.mcp` | Model Context Protocol | | LangChain | `macaw_adapters.langchain` | LangChain (OpenAI, Anthropic, Agents) | ## Examples See the [examples/](examples/) directory for complete working examples: - `examples/openai/` - OpenAI adapter examples - `examples/anthropic/` - Anthropic adapter examples - `examples/langchain/` - LangChain integration examples - `examples/mcp/` - MCP server and client examples ## Console Dev Hub Everything in this repository is also available in the MACAW Console's Dev Hub with interactive features: ``` Console > Dev Hub ├── Quick Start │ └── Download Client SDK (macOS/Linux/Windows, Python 3.9-3.12) and Adapters ├── Tutorials │ └── Role-Based Access Control │ ├── Multi-User SaaS Patterns │ ├── Agent Orchestration │ └── Policy Hierarchies ├── Examples │ ├── OpenAI (drop-in, multi-user, streaming, A2A) │ ├── Anthropic (drop-in, multi-user, streaming, A2A) │ ├── MCP │ │ ├── Simple Invocation │ │ ├── Discovery & Resources │ │ ├── Logging │ │ ├── Progress Tracking │ │ ├── Sampling │ │ ├── Elicitation │ │ └── Roots │ └── LangChain │ ├── Drop-in Agents │ ├── Multi-user Permissions │ ├── Agent Orchestration │ ├── LLM Wrappers (OpenAI, Anthropic) │ └── Memory Integration └── Reference ├── MACAW Client SDK ├── Adapter APIs ├── MAPL Policy Language └── Claims Mapping ``` Access at [console.macawsecurity.ai](https://console.macawsecurity.ai) → Dev Hub tab. ## Research Learn more about the technical foundations of MACAW: - **[Authenticated Workflows: A Systems Approach to Protecting Agentic AI](https://arxiv.org/abs/2602.10465)** - **[Protecting Context and Prompts: Deterministic Security for Non-Deterministic AI](https://arxiv.org/abs/2602.10481)** ## Links - **GitHub**: [github.com/macawsecurity/secureAI](https://github.com/macawsecurity/secureAI) - **Documentation**: [www.macawsecurity.ai/docs](https://www.macawsecurity.ai/docs) - **Console**: [console.macawsecurity.ai](https://console.macawsecurity.ai) - **Support**: help@macawsecurity.com ## License Apache 2.0 - See [LICENSE](LICENSE) for details.
text/markdown
null
MACAW Security <support@macawsecurity.com>
null
null
Apache-2.0
ai, security, openai, claude, langchain, mcp, macaw
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Topic :: Security", "Topic :: Scientific/Engineering :: Artificial Intelligence" ]
[]
null
null
>=3.9
[]
[]
[]
[ "requests>=2.28.0", "cryptography>=41.0.0", "openai>=1.0.0; extra == \"openai\"", "anthropic>=0.18.0; extra == \"claude\"", "langchain>=0.1.0; extra == \"langchain\"", "langchain-openai>=0.0.5; extra == \"langchain\"", "langchain-anthropic>=0.1.0; extra == \"langchain\"", "openai>=1.0.0; extra == \"all\"", "anthropic>=0.18.0; extra == \"all\"", "langchain>=0.1.0; extra == \"all\"", "langchain-openai>=0.0.5; extra == \"all\"", "langchain-anthropic>=0.1.0; extra == \"all\"", "pytest>=7.0.0; extra == \"dev\"", "pytest-asyncio>=0.21.0; extra == \"dev\"", "black>=23.0.0; extra == \"dev\"", "mypy>=1.0.0; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://www.macawsecurity.ai", "Documentation, https://www.macawsecurity.ai/docs", "Repository, https://github.com/macawsecurity/secureAI", "Console, https://console.macawsecurity.ai" ]
twine/6.2.0 CPython/3.9.6
2026-02-21T01:05:34.674469
macaw_adapters-0.6.0.tar.gz
60,487
f7/e3/47342ce7c70ab5e1bd36393517bf634bcf79f761fbedb3596237db818c8f/macaw_adapters-0.6.0.tar.gz
source
sdist
null
false
bed48c3eb3f771dcadc9d7d2c2b522a8
899be9d1d28712d92312d070d875b9c4f33c4f869a2683e6f73c87f7d5045483
f7e347342ce7c70ab5e1bd36393517bf634bcf79f761fbedb3596237db818c8f
null
[ "LICENSE", "NOTICE" ]
230
2.4
aind-dynamic-foraging-models
0.13.1
Generated from aind-library-template
# aind-dynamic-foraging-models [![License](https://img.shields.io/badge/license-MIT-brightgreen)](LICENSE) ![Code Style](https://img.shields.io/badge/code%20style-black-black) [![semantic-release: angular](https://img.shields.io/badge/semantic--release-angular-e10079?logo=semantic-release)](https://github.com/semantic-release/semantic-release) ![Interrogate](https://img.shields.io/badge/interrogate-97.0%25-brightgreen) ![Coverage](https://img.shields.io/badge/coverage-96%25-brightgreen?logo=codecov) ![Python](https://img.shields.io/badge/python->=3.9-blue?logo=python) AIND library for generative (RL) and descriptive (logistic regression) models of dynamic foraging tasks. User documentation available on [readthedocs](https://aind-dynamic-foraging-models.readthedocs.io/). ## Reinforcement Learning (RL) models with Maximum Likelihood Estimation (MLE) fitting ### Overview RL agents that can perform any dynamic foraging task in [aind-behavior-gym](https://github.com/AllenNeuralDynamics/aind-behavior-gym) and can fit behavior using MLE. ![image](https://github.com/user-attachments/assets/b01a208e-b994-4857-a6da-66563bb0916b) ### Code structure ![image](classes_aind_dynamic_foraging_models.svg) - To add more generative models, please subclass [`DynamicForagingAgentMLEBase`](https://github.com/AllenNeuralDynamics/aind-dynamic-foraging-models/blob/11c858f93f67a0699ed23892364f3f51b08eab37/src/aind_dynamic_foraging_models/generative_model/base.py#L25C7-L25C34). ### Implemented foragers <img width="1951" alt="image" src="https://github.com/user-attachments/assets/dacfe875-4e51-492d-a5aa-d3b27ec03e90" /> - [`ForagerQLearning`](https://github.com/AllenNeuralDynamics/aind-dynamic-foraging-models/blob/f9ab39bbdc2cbea350e5a8f11d3f935d6674e08b/src/aind_dynamic_foraging_models/generative_model/forager_q_learning.py): Simple Q-learning agents that incrementally update Q-values. - Available `agent_kwargs`: ```python number_of_learning_rate: Literal[1, 2] = 2, number_of_forget_rate: Literal[0, 1] = 1, choice_kernel: Literal["none", "one_step", "full"] = "none", action_selection: Literal["softmax", "epsilon-greedy"] = "softmax", ``` - [`ForagerLossCounting`](https://github.com/AllenNeuralDynamics/aind-dynamic-foraging-models/blob/f9ab39bbdc2cbea350e5a8f11d3f935d6674e08b/src/aind_dynamic_foraging_models/generative_model/forager_loss_counting.py): Loss counting agents with probabilistic `loss_count_threshold`. - Available `agent_kwargs`: ```python win_stay_lose_switch: Literal[False, True] = False, choice_kernel: Literal["none", "one_step", "full"] = "none", ``` - Action selections ([readthedoc](https://aind-dynamic-foraging-models.readthedocs.io/en/stable/aind_dynamic_foraging_models.generative_model.html#module-aind_dynamic_foraging_models.generative_model.act_functions)) [Here is the full list](https://foraging-behavior-browser.allenneuraldynamics-test.org/RL_model_playground#all-available-foragers) of available foragers: ![image](https://github.com/user-attachments/assets/db2e3b6c-f888-496c-a12b-06e030499165) ![image](https://github.com/user-attachments/assets/4f7b669c-2f0e-49cc-8fb4-7fa948926e2e) ### Usage - [Jupyter notebook](https://github.com/AllenNeuralDynamics/aind-dynamic-foraging-models/blob/main/notebook/demo_RL_agents.ipynb) - See also [these unittest functions](https://github.com/AllenNeuralDynamics/aind-dynamic-foraging-models/tree/main/tests). ### RL model playground Play with the generative models [here](https://foraging-behavior-browser.allenneuraldynamics-test.org/RL_model_playground). ![image](https://github.com/user-attachments/assets/691986b0-114b-437c-8df9-3b7b18f83de9) ## Logistic regression See [this demo notebook.](https://github.com/AllenNeuralDynamics/aind-dynamic-foraging-models/blob/main/notebook/demo_logistic_regression.ipynb) ### Choosing logistic regression models #### Su 2022 ![Untitled](https://hanhou.notion.site/image/https%3A%2F%2Fprod-files-secure.s3.us-west-2.amazonaws.com%2Fe2dab5b7-862e-46a8-aa74-8194ed4315fc%2F54b9718d-6916-48ea-a337-550410a88254%2FUntitled.png?table=block&id=a2db5af7-f2d7-4485-af6c-01a0908546f6&spaceId=e2dab5b7-862e-46a8-aa74-8194ed4315fc&width=1340&userId=&cache=v2) $$ logit(p(c_r)) \sim RewardedChoice+UnrewardedChoice $$ #### Bari 2019 ![Untitled](https://hanhou.notion.site/image/https%3A%2F%2Fprod-files-secure.s3.us-west-2.amazonaws.com%2Fe2dab5b7-862e-46a8-aa74-8194ed4315fc%2F9965a743-89e5-4335-af09-927d96f304e3%2FUntitled.png?table=block&id=1010abe7-4a81-429d-b1b0-5730630e508e&spaceId=e2dab5b7-862e-46a8-aa74-8194ed4315fc&width=1150&userId=&cache=v2) ![Untitled](https://hanhou.notion.site/image/https%3A%2F%2Fprod-files-secure.s3.us-west-2.amazonaws.com%2Fe2dab5b7-862e-46a8-aa74-8194ed4315fc%2Fcb2bbc09-8032-4eb5-8a55-bdadf9f42078%2FUntitled.png?table=block&id=c5cf0499-df10-4ebe-9e81-7eb5e504eede&spaceId=e2dab5b7-862e-46a8-aa74-8194ed4315fc&width=1150&userId=&cache=v2) $$ logit(p(c_r)) \sim RewardedChoice+Choice $$ #### Hattori 2019 ![Untitled](https://hanhou.notion.site/image/https%3A%2F%2Fprod-files-secure.s3.us-west-2.amazonaws.com%2Fe2dab5b7-862e-46a8-aa74-8194ed4315fc%2F44b49866-9f22-45fa-95db-0287a5a9bcfe%2FUntitled.png?table=block&id=20531979-9296-4b51-a41b-bab2e8615c84&spaceId=e2dab5b7-862e-46a8-aa74-8194ed4315fc&width=1340&userId=&cache=v2) $$ logit(p(c_r)) \sim RewardedChoice+UnrewardedChoice+Choice $$ #### Miller 2021 ![Untitled](https://hanhou.notion.site/image/https%3A%2F%2Fprod-files-secure.s3.us-west-2.amazonaws.com%2Fe2dab5b7-862e-46a8-aa74-8194ed4315fc%2F7cd393c3-8c0a-4b63-a4c6-e84a00dff21a%2FUntitled.png?table=block&id=31e3450e-d60f-4c2a-9883-da91a5eaed9b&spaceId=e2dab5b7-862e-46a8-aa74-8194ed4315fc&width=1250&userId=&cache=v2) $$ logit(p(c_r)) \sim Choice + Reward+ Choice*Reward $$ #### Encodings - Ignored trials are removed | choice | reward | Choice | Reward | RewardedChoice | UnrewardedChoice | Choice * Reward | | --- | --- | --- | --- | --- | --- | --- | | L | yes | -1 | 1 | -1 | 0 | -1 | | L | no | -1 | -1 | 0 | -1 | 1 | | R | yes | 1 | 1 | 1 | 0 | 1 | | L | yes | -1 | 1 | -1 | 0 | -1 | | R | no | 1 | -1 | 0 | 1 | -1 | | R | yes | 1 | 1 | 1 | 0 | 1 | | L | no | -1 | -1 | 0 | -1 | 1 | Some observations: 1. $RewardedChoice$ and $UnrewardedChoice$ are orthogonal 2. $Choice = RewardedChoice + UnrewardedChoice$ 3. $Choice * Reward = RewardedChoice - UnrewardedChoice$ #### Comparison | | Su 2022 | Bari 2019 | Hattori 2019 | Miller 2021 | | --- | --- | --- | --- | --- | | Equivalent to | RewC + UnrC | RewC + (RewC + UnrC) | RewC + UnrC + (RewC + UnrC) | (RewC + UnrC) + (RewC - UnrC) + Rew | | Severity of multicollinearity | Not at all | Medium | Severe | Slight | | Interpretation | Like a RL model with different learning rates on reward and unrewarded trials. | Like a RL model that only updates on rewarded trials, plus a choice kernel (tendency to repeat previous choices). | Like a RL model that has different learning rates on reward and unrewarded trials, plus a choice kernel (the full RL model from the same paper). | Like a RL model that has symmetric learning rates for rewarded and unrewarded trials, plus a choice kernel. However, the $Reward $ term seems to be a strawman assumption, as it means “if I get reward on any side, I’ll choose the right side more”, which doesn’t make much sense. | | Conclusion | Probably the best | Okay | Not good due to the severe multicollinearity | Good | ### Regularization and optimization The choice of optimizer depends on the penality term, as listed [here](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html#sklearn.linear_model.LogisticRegression). - `lbfgs` - [`l2`, None] - `liblinear` - [`l1`, `l2`] - `newton-cg` - [`l2`, None] - `newton-cholesky` - [`l2`, None] - `sag` - [`l2`, None] - `saga` - [`elasticnet`, `l1`, `l2`, None] ## See also - Foraging model simulation, model recovery, etc.: https://github.com/hanhou/Dynamic-Foraging ## Installation To install the package (logistic regression only), run ```bash pip install aind-dynamic-foraging-models ``` To install the package with RL models, run ```bash pip install aind-dynamic-foraging-models[rl] ``` To develop the code, clone the repo to your local machine, and run ```bash pip install -e .[dev] ``` ## Contributing ### Linters and testing There are several libraries used to run linters, check documentation, and run tests. - Please test your changes using the **coverage** library, which will run the tests and log a coverage report: ```bash coverage run -m unittest discover && coverage report ``` - Use **interrogate** to check that modules, methods, etc. have been documented thoroughly: ```bash interrogate . ``` - Use **flake8** to check that code is up to standards (no unused imports, etc.): ```bash flake8 . ``` - Use **black** to automatically format the code into PEP standards: ```bash black . ``` - Use **isort** to automatically sort import statements: ```bash isort . ``` ### Pull requests For internal members, please create a branch. For external members, please fork the repository and open a pull request from the fork. We'll primarily use [Angular](https://github.com/angular/angular/blob/main/CONTRIBUTING.md#commit) style for commit messages. Roughly, they should follow the pattern: ```text <type>(<scope>): <short summary> ``` where scope (optional) describes the packages affected by the code changes and type (mandatory) is one of: - **build**: Changes that affect build tools or external dependencies (example scopes: pyproject.toml, setup.py) - **ci**: Changes to our CI configuration files and scripts (examples: .github/workflows/ci.yml) - **docs**: Documentation only changes - **feat**: A new feature - **fix**: A bugfix - **perf**: A code change that improves performance - **refactor**: A code change that neither fixes a bug nor adds a feature - **test**: Adding missing tests or correcting existing tests ### Semantic Release The table below, from [semantic release](https://github.com/semantic-release/semantic-release), shows which commit message gets you which release type when `semantic-release` runs (using the default configuration): | Commit message | Release type | | ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ | --------------------------------------------------------------------------------------------------------------- | | `fix(pencil): stop graphite breaking when too much pressure applied` | ~~Patch~~ Fix Release, Default release | | `feat(pencil): add 'graphiteWidth' option` | ~~Minor~~ Feature Release | | `perf(pencil): remove graphiteWidth option`<br><br>`BREAKING CHANGE: The graphiteWidth option has been removed.`<br>`The default graphite width of 10mm is always used for performance reasons.` | ~~Major~~ Breaking Release <br /> (Note that the `BREAKING CHANGE: ` token must be in the footer of the commit) | ### Documentation To generate the rst files source files for documentation, run ```bash sphinx-apidoc -o doc_template/source/ src ``` Then to create the documentation HTML files, run ```bash sphinx-build -b html doc_template/source/ doc_template/build/html ``` More info on sphinx installation can be found [here](https://www.sphinx-doc.org/en/master/usage/installation.html).
text/markdown
Allen Institute for Neural Dynamics
null
null
null
MIT
null
[ "Programming Language :: Python :: 3" ]
[]
null
null
>=3.9
[]
[]
[]
[ "numpy", "matplotlib", "scikit-learn", "scipy", "pydantic", "black; extra == \"dev\"", "coverage; extra == \"dev\"", "flake8; extra == \"dev\"", "interrogate; extra == \"dev\"", "isort; extra == \"dev\"", "Sphinx; extra == \"dev\"", "furo; extra == \"dev\"", "aind-behavior-gym; extra == \"dev\"", "aind_dynamic_foraging_basic_analysis; extra == \"dev\"", "aind-behavior-gym; extra == \"rl\"", "aind_dynamic_foraging_basic_analysis; extra == \"rl\"" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-21T01:05:25.913338
aind_dynamic_foraging_models-0.13.1.tar.gz
6,648,444
eb/47/a7b35997f9767cc4911ce65db39a06492192d338175487916d98b70b1640/aind_dynamic_foraging_models-0.13.1.tar.gz
source
sdist
null
false
4d9f28fa43b3f082d6b5b5db24d36e26
260e8cacb8b55b43f12b2443f876ced4d1cde364f9938199c1b2466425cd2419
eb47a7b35997f9767cc4911ce65db39a06492192d338175487916d98b70b1640
null
[ "LICENSE" ]
230
2.4
langflow-nightly
1.8.0.dev57
A Python package with a built-in web application
<!-- markdownlint-disable MD030 --> <picture> <source media="(prefers-color-scheme: dark)" srcset="./docs/static/img/langflow-logo-color-blue-bg.svg"> <img src="./docs/static/img/langflow-logo-color-black-solid.svg" alt="Langflow logo"> </picture> [![Release Notes](https://img.shields.io/github/release/langflow-ai/langflow?style=flat-square)](https://github.com/langflow-ai/langflow/releases) [![PyPI - License](https://img.shields.io/badge/license-MIT-orange)](https://opensource.org/licenses/MIT) [![PyPI - Downloads](https://img.shields.io/pypi/dm/langflow?style=flat-square)](https://pypistats.org/packages/langflow) [![Twitter](https://img.shields.io/twitter/url/https/twitter.com/langflow-ai.svg?style=social&label=Follow%20%40Langflow)](https://twitter.com/langflow_ai) [![YouTube Channel](https://img.shields.io/youtube/channel/subscribers/UCn2bInQrjdDYKEEmbpwblLQ?label=Subscribe)](https://www.youtube.com/@Langflow) [![Discord Server](https://img.shields.io/discord/1116803230643527710?logo=discord&style=social&label=Join)](https://discord.gg/EqksyE2EX9) [![Ask DeepWiki](https://deepwiki.com/badge.svg)](https://deepwiki.com/langflow-ai/langflow) [Langflow](https://langflow.org) is a powerful platform for building and deploying AI-powered agents and workflows. It provides developers with both a visual authoring experience and built-in API and MCP servers that turn every workflow into a tool that can be integrated into applications built on any framework or stack. Langflow comes with batteries included and supports all major LLMs, vector databases and a growing library of AI tools. ## ✨ Highlight features - **Visual builder interface** to quickly get started and iterate. - **Source code access** lets you customize any component using Python. - **Interactive playground** to immediately test and refine your flows with step-by-step control. - **Multi-agent orchestration** with conversation management and retrieval. - **Deploy as an API** or export as JSON for Python apps. - **Deploy as an MCP server** and turn your flows into tools for MCP clients. - **Observability** with LangSmith, LangFuse and other integrations. - **Enterprise-ready** security and scalability. ## 🖥️ Langflow Desktop Langflow Desktop is the easiest way to get started with Langflow. All dependencies are included, so you don't need to manage Python environments or install packages manually. Available for Windows and macOS. [📥 Download Langflow Desktop](https://www.langflow.org/desktop) ## ⚡️ Quickstart ### Install locally (recommended) Requires Python 3.10–3.13 and [uv](https://docs.astral.sh/uv/getting-started/installation/) (recommended package manager). #### Install From a fresh directory, run: ```shell uv pip install langflow -U ``` The latest Langflow package is installed. For more information, see [Install and run the Langflow OSS Python package](https://docs.langflow.org/get-started-installation#install-and-run-the-langflow-oss-python-package). #### Run To start Langflow, run: ```shell uv run langflow run ``` Langflow starts at http://127.0.0.1:7860. That's it! You're ready to build with Langflow! 🎉 ## 📦 Other install options ### Run from source If you've cloned this repository and want to contribute, run this command from the repository root: ```shell make run_cli ``` For more information, see [DEVELOPMENT.md](./DEVELOPMENT.md). ### Docker Start a Langflow container with default settings: ```shell docker run -p 7860:7860 langflowai/langflow:latest ``` Langflow is available at http://localhost:7860/. For configuration options, see the [Docker deployment guide](https://docs.langflow.org/deployment-docker). > [!CAUTION] > - Users must update to Langflow >= 1.7.1 to protect against [CVE-2025-68477](https://github.com/langflow-ai/langflow/security/advisories/GHSA-5993-7p27-66g5) and [CVE-2025-68478](https://github.com/langflow-ai/langflow/security/advisories/GHSA-f43r-cc68-gpx4). > - Langflow version 1.7.0 has a critical bug where persisted state (flows, projects, and global variables) cannot be found when upgrading. Version 1.7.0 was yanked and replaced with version 1.7.1, which includes a fix for this bug. **DO NOT** upgrade to version 1.7.0. Instead, upgrade directly to version 1.7.1. > - Langflow versions 1.6.0 through 1.6.3 have a critical bug where `.env` files are not read, potentially causing security vulnerabilities. **DO NOT** upgrade to these versions if you use `.env` files for configuration. Instead, upgrade to 1.6.4, which includes a fix for this bug. > - Windows users of Langflow Desktop should **not** use the in-app update feature to upgrade to Langflow version 1.6.0. For upgrade instructions, see [Windows Desktop update issue](https://docs.langflow.org/release-notes#windows-desktop-update-issue). > - Users must update to Langflow >= 1.3 to protect against [CVE-2025-3248](https://nvd.nist.gov/vuln/detail/CVE-2025-3248) > - Users must update to Langflow >= 1.5.1 to protect against [CVE-2025-57760](https://github.com/langflow-ai/langflow/security/advisories/GHSA-4gv9-mp8m-592r) > > For security information, see our [Security Policy](./SECURITY.md) and [Security Advisories](https://github.com/langflow-ai/langflow/security/advisories). ## 🚀 Deployment Langflow is completely open source and you can deploy it to all major deployment clouds. To learn how to deploy Langflow, see our [Langflow deployment guides](https://docs.langflow.org/deployment-overview). ## ⭐ Stay up-to-date Star Langflow on GitHub to be instantly notified of new releases. ![Star Langflow](https://github.com/user-attachments/assets/03168b17-a11d-4b2a-b0f7-c1cce69e5a2c) ## 👋 Contribute We welcome contributions from developers of all levels. If you'd like to contribute, please check our [contributing guidelines](./CONTRIBUTING.md) and help make Langflow more accessible. --- [![Star History Chart](https://api.star-history.com/svg?repos=langflow-ai/langflow&type=Timeline)](https://star-history.com/#langflow-ai/langflow&Date) ## ❤️ Contributors [![langflow contributors](https://contrib.rocks/image?repo=langflow-ai/langflow)](https://github.com/langflow-ai/langflow/graphs/contributors)
text/markdown
null
null
null
Carlos Coelho <carlos@langflow.org>, Cristhian Zanforlin <cristhian.lousa@gmail.com>, Gabriel Almeida <gabriel@langflow.org>, Lucas Eduoli <lucaseduoli@gmail.com>, Otávio Anovazzi <otavio2204@gmail.com>, Rodrigo Nader <rodrigo@langflow.org>, Italo dos Anjos <italojohnnydosanjos@gmail.com>
null
gpt, gui, langchain, nlp, openai
[]
[]
null
null
<3.14,>=3.10
[]
[]
[]
[ "langflow-base-nightly[complete]==1.8.0.dev57", "webrtcvad>=2.0.10; extra == \"audio\"", "cassio>=0.1.7; extra == \"cassio\"", "clickhouse-connect==0.7.19; extra == \"clickhouse-connect\"", "couchbase>=4.2.1; extra == \"couchbase\"", "langchain-docling>=1.1.0; extra == \"docling\"", "ocrmac>=1.0.0; sys_platform == \"darwin\" and extra == \"docling\"", "rapidocr-onnxruntime>=1.4.4; extra == \"docling\"", "tesserocr>=2.8.0; extra == \"docling\"", "ctransformers>=0.2.10; extra == \"local\"", "llama-cpp-python~=0.2.0; extra == \"local\"", "sentence-transformers>=2.3.1; extra == \"local\"", "nv-ingest-api<26.0.0,==25.6.2; python_version >= \"3.12\" and extra == \"nv-ingest\"", "nv-ingest-client<26.0.0,==25.6.3; python_version >= \"3.12\" and extra == \"nv-ingest\"", "sqlalchemy[postgresql-psycopg2binary]<3.0.0,>=2.0.38; extra == \"postgresql\"", "sqlalchemy[postgresql-psycopg]<3.0.0,>=2.0.38; extra == \"postgresql\"" ]
[]
[]
[]
[ "Repository, https://github.com/langflow-ai/langflow", "Documentation, https://docs.langflow.org" ]
uv/0.10.4 {"installer":{"name":"uv","version":"0.10.4","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
2026-02-21T01:05:10.424083
langflow_nightly-1.8.0.dev57-py3-none-any.whl
6,259
86/c5/7ca1fe07965a6e50ef18f897f7d6a4b25c67710f635756e43165e34e6cee/langflow_nightly-1.8.0.dev57-py3-none-any.whl
py3
bdist_wheel
null
false
3fd27b244aa895992a610b8a166c255f
528ad67d737be82d3d08f6c6b15516d78ee06db1bd8331b2feb2d9f37dbc0797
86c57ca1fe07965a6e50ef18f897f7d6a4b25c67710f635756e43165e34e6cee
MIT
[ "LICENSE" ]
68
2.4
lfx-nightly
0.3.0.dev57
Langflow Executor - A lightweight CLI tool for executing and serving Langflow AI flows
# lfx - Langflow Executor lfx is a command-line tool for running Langflow workflows. It provides two main commands: `serve` and `run`. ## Installation ### From PyPI (recommended) ```bash # Install globally uv pip install lfx # Or run without installing using uvx uvx lfx serve my_flow.json uvx lfx run my_flow.json "input" ``` ### From source (development) ```bash # Clone and run in workspace git clone https://github.com/langflow-ai/langflow cd langflow/src/lfx uv run lfx serve my_flow.json ``` ## Key Features ### Pluggable Services lfx supports a pluggable service architecture that allows you to customize and extend its behavior. You can replace built-in services (storage, telemetry, tracing, etc.) with your own implementations or use Langflow's full-featured services. 📖 **See [PLUGGABLE_SERVICES.md](./PLUGGABLE_SERVICES.md) for details** including: - Quick start guides for CLI users, library developers, and plugin authors - Service registration via config files, decorators, and entry points - Creating custom service implementations with dependency injection - Using full-featured Langflow services in lfx - Troubleshooting and migration guides ### Flattened Component Access lfx now supports simplified component imports for better developer experience: **Before (old import style):** ```python from lfx.components.agents.agent import AgentComponent from lfx.components.data.url import URLComponent from lfx.components.input_output import ChatInput, ChatOutput ``` **Now (new flattened style):** ```python from lfx import components as cp # Direct access to all components chat_input = cp.ChatInput() agent = cp.AgentComponent() url_component = cp.URLComponent() chat_output = cp.ChatOutput() ``` **Benefits:** - **Simpler imports**: One import line instead of multiple deep imports - **Better discovery**: All components accessible via `cp.ComponentName` - **Helpful error messages**: Clear guidance when dependencies are missing - **Backward compatible**: Traditional imports still work ## Commands ### `lfx serve` - Run flows as an API Serve a Langflow workflow as a REST API. **Important:** You must set the `LANGFLOW_API_KEY` environment variable before running the serve command. ```bash export LANGFLOW_API_KEY=your-secret-key uv run lfx serve my_flow.json --port 8000 ``` This creates a FastAPI server with your flow available at `/flows/{flow_id}/run`. The actual flow ID will be displayed when the server starts. **Options:** - `--host, -h`: Host to bind server (default: 127.0.0.1) - `--port, -p`: Port to bind server (default: 8000) - `--verbose, -v`: Show diagnostic output - `--env-file`: Path to .env file - `--log-level`: Set logging level (debug, info, warning, error, critical) - `--check-variables/--no-check-variables`: Check global variables for environment compatibility (default: check) **Example:** ```bash # Set API key (required) export LANGFLOW_API_KEY=your-secret-key # Start server uv run lfx serve simple_chat.json --host 0.0.0.0 --port 8000 # The server will display the flow ID, e.g.: # Flow ID: af9edd65-6393-58e2-9ae5-d5f012e714f4 # Call API using the displayed flow ID curl -X POST http://localhost:8000/flows/af9edd65-6393-58e2-9ae5-d5f012e714f4/run \ -H "Content-Type: application/json" \ -H "x-api-key: your-secret-key" \ -d '{"input_value": "Hello, world!"}' ``` ### `lfx run` - Run flows directly Execute a Langflow workflow and get results immediately. ```bash uv run lfx run my_flow.json "What is AI?" ``` **Options:** - `--format, -f`: Output format (json, text, message, result) (default: json) - `--verbose`: Show diagnostic output - `--input-value`: Input value to pass to the graph (alternative to positional argument) - `--flow-json`: Inline JSON flow content as a string - `--stdin`: Read JSON flow from stdin - `--check-variables/--no-check-variables`: Check global variables for environment compatibility (default: check) **Examples:** ```bash # Basic execution uv run lfx run simple_chat.json "Tell me a joke" # JSON output (default) uv run lfx run simple_chat.json "input text" --format json # Text output only uv run lfx run simple_chat.json "Hello" --format text # Using --input-value flag uv run lfx run simple_chat.json --input-value "Hello world" # From stdin (requires --input-value for input) echo '{"data": {"nodes": [...], "edges": [...]}}' | uv run lfx run --stdin --input-value "Your message" # Inline JSON uv run lfx run --flow-json '{"data": {"nodes": [...], "edges": [...]}}' --input-value "Test" ``` ### Complete Agent Example Here's a step-by-step example of creating and running an agent workflow with dependencies: **Step 1: Create the agent script** Create a file called `simple_agent.py`: ```python """A simple agent flow example for Langflow. This script demonstrates how to set up a conversational agent using Langflow's Agent component with web search capabilities. Features: - Uses the new flattened component access (cp.AgentComponent instead of deep imports) - Configures logging to 'langflow.log' at INFO level - Creates an agent with OpenAI GPT model - Provides web search tools via URLComponent - Connects ChatInput → Agent → ChatOutput - Uses async get_graph() function for proper async handling Usage: uv run lfx run simple_agent.py "How are you?" """ import os from pathlib import Path # Using the new flattened component access from lfx import components as cp from lfx.graph import Graph from lfx.log.logger import LogConfig async def get_graph() -> Graph: """Create and return the graph with async component initialization. This function properly handles async component initialization without blocking the module loading process. The script loader will detect this async function and handle it appropriately. Returns: Graph: The configured graph with ChatInput → Agent → ChatOutput flow """ log_config = LogConfig( log_level="INFO", log_file=Path("langflow.log"), ) # Showcase the new flattened component access - no need for deep imports! chat_input = cp.ChatInput() agent = cp.AgentComponent() # Use URLComponent for web search capabilities url_component = cp.URLComponent() tools = await url_component.to_toolkit() agent.set( model_name="gpt-4.1-mini", agent_llm="OpenAI", api_key=os.getenv("OPENAI_API_KEY"), input_value=chat_input.message_response, tools=tools, ) chat_output = cp.ChatOutput().set(input_value=agent.message_response) return Graph(chat_input, chat_output, log_config=log_config) ``` **Step 2: Install dependencies** ```bash # Install lfx (if not already installed) uv pip install lfx # Install additional dependencies required for the agent uv pip install 'langchain-core>=0.3.0,<1.0.0' \ 'langchain-openai>=0.3.0,<1.0.0' \ 'langchain-community>=0.3.0,<1.0.0' \ beautifulsoup4 lxml ``` **Step 3: Set up environment** ```bash # Set your OpenAI API key export OPENAI_API_KEY=your-openai-api-key-here ``` **Step 4: Run the agent** ```bash # Run with verbose output to see detailed execution uv run lfx run simple_agent.py "How are you?" --verbose # Run with different questions uv run lfx run simple_agent.py "What's the weather like today?" uv run lfx run simple_agent.py "Search for the latest news about AI" ``` This creates an intelligent agent that can: - Answer questions using the GPT model - Search the web for current information - Process and respond to natural language queries The `--verbose` flag shows detailed execution information including timing and component details. ## Input Sources Both commands support multiple input sources: - **File path**: `uv run lfx serve my_flow.json` - **Inline JSON**: `uv run lfx serve --flow-json '{"data": {"nodes": [...], "edges": [...]}}'` - **Stdin**: `uv run lfx serve --stdin` ## Development ```bash # Install development dependencies make dev # Run tests make test # Format code make format ``` ## License MIT License. See [LICENSE](../../LICENSE) for details.
text/markdown
null
Gabriel Luiz Freitas Almeida <gabriel@langflow.org>
null
null
null
null
[]
[]
null
null
<3.14,>=3.10
[]
[]
[]
[ "ag-ui-protocol>=0.1.10", "aiofile<4.0.0,>=3.8.0", "aiofiles<25.0.0,>=24.1.0", "asyncer<1.0.0,>=0.0.8", "cachetools>=6.0.0", "chardet<6.0.0,>=5.2.0", "cryptography>=43.0.0", "defusedxml<1.0.0,>=0.7.1", "docstring-parser<1.0.0,>=0.16", "emoji<3.0.0,>=2.14.1", "fastapi<1.0.0,>=0.115.13", "filelock<4.0.0,>=3.20.1", "httpx[http2]<1.0.0,>=0.24.0", "json-repair<1.0.0,>=0.30.3", "langchain-core<1.0.0,>=0.3.81", "langchain~=0.3.23", "loguru<1.0.0,>=0.7.3", "markitdown<2.0.0,>=0.1.4", "nanoid<3.0.0,>=2.0.0", "networkx<4.0.0,>=3.4.2", "orjson<4.0.0,>=3.10.15", "pandas<3.0.0,>=2.0.0", "passlib<2.0.0,>=1.7.4", "pillow<13.0.0,>=10.0.0", "platformdirs<5.0.0,>=4.3.8", "pydantic-settings<3.0.0,>=2.10.1", "pydantic<3.0.0,>=2.0.0", "pypdf<7.0.0,>=6.4.0", "python-dotenv<2.0.0,>=1.0.0", "rich<14.0.0,>=13.0.0", "setuptools<81.0.0,>=80.0.0", "structlog<26.0.0,>=25.4.0", "tomli<3.0.0,>=2.2.1", "typer<1.0.0,>=0.16.0", "typing-extensions<5.0.0,>=4.14.0", "uvicorn<1.0.0,>=0.34.3", "validators<1.0.0,>=0.34.0", "wheel<1.0.0,>=0.46.2" ]
[]
[]
[]
[]
uv/0.10.4 {"installer":{"name":"uv","version":"0.10.4","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
2026-02-21T01:04:33.575924
lfx_nightly-0.3.0.dev57-py3-none-any.whl
1,815,643
37/95/96e165933d9aa63c372105a0d63b7b821f4a13b3ab7f5f372ae4a7839bdb/lfx_nightly-0.3.0.dev57-py3-none-any.whl
py3
bdist_wheel
null
false
ed282f1b3bbf5d13ad603a3b764fd871
da0a6d5cea5048e9de2edaf094e842a20ad9814276a295b933677022a54e4f6e
379596e165933d9aa63c372105a0d63b7b821f4a13b3ab7f5f372ae4a7839bdb
null
[]
75
2.4
pyTMD
3.0.3
Python-based tidal prediction software for estimating ocean, load, solid Earth and pole tides
# pyTMD Python-based tidal prediction software for estimating ocean, load, solid Earth and pole tides ## About <table> <tr> <td><b>Version:</b></td> <td> <a href="https://pypi.python.org/pypi/pyTMD/" alt="PyPI"><img src="https://img.shields.io/pypi/v/pyTMD.svg"></a> <a href="https://anaconda.org/conda-forge/pytmd" alt="conda-forge"><img src="https://img.shields.io/conda/vn/conda-forge/pytmd"></a> <a href="https://github.com/pyTMD/pyTMD/releases/latest" alt="commits-since"><img src="https://img.shields.io/github/commits-since/pyTMD/pyTMD/latest"></a> </td> </tr> <tr> <td><b>Citation:</b></td> <td> <a href="https://doi.org/10.21105/joss.08566" alt="JOSS"><img src="https://joss.theoj.org/papers/10.21105/joss.08566/status.svg"></a> <a href="https://doi.org/10.5281/zenodo.5555395" alt="zenodo"><img src="https://zenodo.org/badge/DOI/10.5281/zenodo.5555395.svg"></a> </td> </tr> <tr> <td><b>Tests:</b></td> <td> <a href="https://pytmd.readthedocs.io/en/latest/?badge=latest" alt="Documentation Status"><img src="https://readthedocs.org/projects/pytmd/badge/?version=latest"></a> <a href="https://github.com/pyTMD/pyTMD/actions/workflows/python-request.yml" alt="Build"><img src="https://github.com/pyTMD/pyTMD/actions/workflows/python-request.yml/badge.svg"></a> <a href="https://github.com/pyTMD/pyTMD/actions/workflows/ruff-format.yml" alt="Ruff"><img src="https://github.com/pyTMD/pyTMD/actions/workflows/ruff-format.yml/badge.svg"></a> </td> </tr> <tr> <td><b>Data:</b></td> <td> <a href="https://doi.org/10.5281/zenodo.18091740" alt="zenodo"><img src="https://img.shields.io/badge/zenodo-pyTMD_test_data-2f6fa7.svg?logo=data:image/svg%2bxml;base64,PHN2ZyBpZD0iU3ZnanNTdmcxMDIxIiB3aWR0aD0iMjg4IiBoZWlnaHQ9IjI4OCIgeG1sbnM9Imh0dHA6Ly93d3cudzMub3JnLzIwMDAvc3ZnIiB2ZXJzaW9uPSIxLjEiIHhtbG5zOnhsaW5rPSJodHRwOi8vd3d3LnczLm9yZy8xOTk5L3hsaW5rIiB4bWxuczpzdmdqcz0iaHR0cDovL3N2Z2pzLmNvbS9zdmdqcyI+PGRlZnMgaWQ9IlN2Z2pzRGVmczEwMjIiPjwvZGVmcz48ZyBpZD0iU3ZnanNHMTAyMyI+PHN2ZyB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciIGVuYWJsZS1iYWNrZ3JvdW5kPSJuZXcgMCAwIDIyMCA4MCIgdmlld0JveD0iMCAwIDUxLjA0NiA1MS4wNDYiIHdpZHRoPSIyODgiIGhlaWdodD0iMjg4Ij48cGF0aCBmaWxsPSIjZmZmZmZmIiBkPSJtIDI4LjMyNCwyMC4wNDQgYyAtMC4wNDMsLTAuMTA2IC0wLjA4NCwtMC4yMTQgLTAuMTMxLC0wLjMyIC0wLjcwNywtMS42MDIgLTEuNjU2LC0yLjk5NyAtMi44NDgsLTQuMTkgLTEuMTg4LC0xLjE4NyAtMi41ODIsLTIuMTI1IC00LjE4NCwtMi44MDUgLTEuNjA1LC0wLjY3OCAtMy4zMDksLTEuMDIgLTUuMTA0LC0xLjAyIC0xLjg1LDAgLTMuNTY0LDAuMzQyIC01LjEzNywxLjAyIC0xLjQ2NywwLjYyOCAtMi43NjQsMS40ODggLTMuOTEsMi41NTIgViAxNC44NCBjIDAsLTEuNTU3IC0xLjI2MiwtMi44MjIgLTIuODIsLTIuODIyIGggLTE5Ljc3NSBjIC0xLjU1NywwIC0yLjgyLDEuMjY1IC0yLjgyLDIuODIyIDAsMS41NTkgMS4yNjQsMi44MiAyLjgyLDIuODIgaCAxNS41NDEgbCAtMTguMjMsMjQuNTQ2IGMgLTAuMzYyLDAuNDg3IC0wLjU1NywxLjA3NyAtMC41NTcsMS42ODIgdiAxLjg0MSBjIDAsMS41NTggMS4yNjQsMi44MjIgMi44MjIsMi44MjIgSCA1LjAzOCBjIDEuNDg4LDAgMi43MDUsLTEuMTUzIDIuODEyLC0yLjYxNCAwLjkzMiwwLjc0MyAxLjk2NywxLjM2NCAzLjEwOSwxLjg0OCAxLjYwNSwwLjY4NCAzLjI5OSwxLjAyMSA1LjEwMiwxLjAyMSAyLjcyMywwIDUuMTUsLTAuNzI2IDcuMjg3LC0yLjE4NyAxLjcyNywtMS4xNzYgMy4wOTIsLTIuNjM5IDQuMDg0LC00LjM4OSAwLjgzMjc5OSwtMS40NzIwOTQgMS40MTgyODQsLTIuNjMzMzUyIDEuMjIxODg5LC0zLjcyOTE4MiAtMC4xNzMwMDMsLTAuOTY1MzE4IC0wLjY5NDkxNCwtMS45NDY0MTkgLTIuMzI2ODY1LC0yLjM3ODM1OCAtMC41OCwwIC0xLjM3NjAyNCwwLjE3NDU0IC0xLjgzMzAyNCwwLjQ5MjU0IC0wLjQ2MywwLjMxNiAtMC43OTMsMC43NDQgLTAuOTgyLDEuMjc1IGwgLTAuNDUzLDAuOTMgYyAtMC42MzEsMS4zNjUgLTEuNTY2LDIuNDQzIC0yLjgwOSwzLjI0NCAtMS4yMzgsMC44MDMgLTIuNjMzLDEuMjAxIC00LjE4OCwxLjIwMSAtMS4wMjMsMCAtMi4wMDQsLTAuMTkxIC0yLjk1NSwtMC41NzkgLTAuOTQxLC0wLjM5IC0xLjc1OCwtMC45MzUgLTIuNDM5LC0xLjY0IEMgOS45ODYsNDAuMzQzIDkuNDQxLDM5LjUyNiA5LjAyNywzOC42MDMgOC42MTcsMzcuNjc5IDguNDEsMzYuNzEgOC40MSwzNS42ODcgdiAtMi40NzYgaCAxNy43MTUgYyAwLDAgMS41MTc3NzQsLTAuMTU0NjYgMi4xODMzNzUsLTAuNzcwNjcyIDAuOTU4NDk2LC0wLjg4NzA4NSAwLjg2NDYyMiwtMi4xNTAzOCAwLjg2NDYyMiwtMi4xNTAzOCAwLDAgLTAuMDQzNTQsLTUuMDY2ODM0IC0wLjMzODM3NiwtNy41NzgxNTQgQyAyOC43MjkwNDgsMjEuODEyNTYzIDI4LjMyNCwyMC4wNDQgMjguMzI0LDIwLjA0NCBaIE0gLTExLjc2Nyw0Mi45MSAyLjk5MSwyMy4wMzYgQyAyLjkxMywyMy42MjMgMi44NywyNC4yMiAyLjg3LDI0LjgyNyB2IDEwLjg2IGMgMCwxLjc5OSAwLjM1LDMuNDk4IDEuMDU5LDUuMTA0IDAuMzI4LDAuNzUyIDAuNzE5LDEuNDU4IDEuMTU2LDIuMTE5IC0wLjAxNiwwIC0wLjAzMSwtMTBlLTQgLTAuMDQ3LC0xMGUtNCBIIC0xMS43NjcgWiBNIDIzLjcxLDI3LjY2NyBIIDguNDA5IHYgLTIuODQxIGMgMCwtMS4wMTUgMC4xODksLTEuOTkgMC41OCwtMi45MTIgMC4zOTEsLTAuOTIyIDAuOTM2LC0xLjc0IDEuNjQ1LC0yLjQ0NCAwLjY5NywtMC43MDMgMS41MTYsLTEuMjQ5IDIuNDM4LC0xLjY0MSAwLjkyMiwtMC4zODggMS45MiwtMC41ODEgMi45OSwtMC41ODEgMS4wMiwwIDIuMDAyLDAuMTkzIDIuOTQ5LDAuNTgxIDAuOTQ5LDAuMzkzIDEuNzY0LDAuOTM4IDIuNDQxLDEuNjQxIDAuNjgyLDAuNzA0IDEuMjI1LDEuNTIxIDEuNjQxLDIuNDQ0IDAuNDE0LDAuOTIyIDAuNjE3LDEuODk2IDAuNjE3LDIuOTEyIHoiIHRyYW5zZm9ybT0idHJhbnNsYXRlKDIwLjM1IC00LjczNSkiIGNsYXNzPSJjb2xvcmZmZiBzdmdTaGFwZSI+PC9wYXRoPjwvc3ZnPjwvZz48L3N2Zz4="></a> <a href="https://doi.org/10.6084/m9.figshare.30260326" alt="figshare"><img src="https://img.shields.io/badge/figshare-pyTMD_test_data-a60845?logo=figshare"></a> </td> </tr> <tr> <td><b>License:</b></td> <td> <a href="https://github.com/pyTMD/pyTMD/blob/main/LICENSE" alt="License"><img src="https://img.shields.io/github/license/pyTMD/pyTMD"></a> </td> </tr> </table> For more information: see the documentation at [pytmd.readthedocs.io](https://pytmd.readthedocs.io/) ## Installation From PyPI: ```bash python3 -m pip install pyTMD ``` To include all optional dependencies: ```bash python3 -m pip install pyTMD[all] ``` Using `conda` or `mamba` from conda-forge: ```bash conda install -c conda-forge pytmd ``` ```bash mamba install -c conda-forge pytmd ``` Development version from GitHub: ```bash python3 -m pip install git+https://github.com/pyTMD/pyTMD.git ``` ### Running with Pixi Alternatively, you can use [Pixi](https://pixi.sh/) for a streamlined workspace environment: 1. Install Pixi following the [installation instructions](https://pixi.sh/latest/#installation) 2. Clone the project repository: ```bash git clone https://github.com/pyTMD/pyTMD.git ``` 3. Move into the `pyTMD` directory ```bash cd pyTMD ``` 4. Install dependencies and start JupyterLab: ```bash pixi run start ``` This will automatically create the environment, install all dependencies, and launch JupyterLab in the [notebooks](./doc/source/notebooks/) directory. ## Dependencies - [h5netcdf: Pythonic interface to netCDF4 via h5py](https://h5netcdf.org/) - [lxml: processing XML and HTML in Python](https://pypi.python.org/pypi/lxml) - [numpy: Scientific Computing Tools For Python](https://www.numpy.org) - [platformdirs: Python module for determining platform-specific directories](https://pypi.org/project/platformdirs/) - [pyproj: Python interface to PROJ library](https://pypi.org/project/pyproj/) - [scipy: Scientific Tools for Python](https://www.scipy.org/) - [timescale: Python tools for time and astronomical calculations](https://pypi.org/project/timescale/) - [xarray: N-D labeled arrays and datasets in Python](https://docs.xarray.dev/en/stable/) ## References > T. C. Sutterley, S. L. Howard, L. Padman, and M. R. Siegfried, > "pyTMD: Python-based tidal prediction software". *Journal of Open Source Software*, > 10(116), 8566, (2025). [doi: 10.21105/joss.08566](https://doi.org/10.21105/joss.08566) > > T. C. Sutterley, T. Markus, T. A. Neumann, M. R. van den Broeke, J. M. van Wessem, and S. R. M. Ligtenberg, > "Antarctic ice shelf thickness change from multimission lidar mapping", *The Cryosphere*, > 13, 1801-1817, (2019). [doi: 10.5194/tc-13-1801-2019](https://doi.org/10.5194/tc-13-1801-2019) > > L. Padman, M. R. Siegfried, and H. A. Fricker, > "Ocean Tide Influences on the Antarctic and Greenland Ice Sheets", *Reviews of Geophysics*, > 56, 142-184, (2018). [doi: 10.1002/2016RG000546](https://doi.org/10.1002/2016RG000546) ## Download The program homepage is: <https://github.com/pyTMD/pyTMD> A zip archive of the latest version is available directly at: <https://github.com/pyTMD/pyTMD/archive/main.zip> ## Alternative Software perth5 from NASA Goddard Space Flight Center: <https://codeberg.org/rray/perth5> Matlab Tide Model Driver from Earth & Space Research: <https://github.com/EarthAndSpaceResearch/TMD_Matlab_Toolbox_v2.5> Fortran OSU Tidal Prediction Software: <https://www.tpxo.net/otps> ## Disclaimer This package includes software developed at NASA Goddard Space Flight Center (GSFC) and the University of Washington Applied Physics Laboratory (UW-APL). It is not sponsored or maintained by the Universities Space Research Association (USRA), AVISO or NASA. The software is provided here for your convenience but *with no guarantees whatsoever*. It should not be used for coastal navigation or any application that may risk life or property. ## Contributing This project contains work and contributions from the [scientific community](./CONTRIBUTORS.md). If you would like to contribute to the project, please have a look at the [contribution guidelines](./doc/source/getting_started/Contributing.rst), [open issues](https://github.com/pyTMD/pyTMD/issues) and [discussions board](https://github.com/pyTMD/pyTMD/discussions). ## Credits The Tidal Model Driver ([TMD](https://github.com/EarthAndSpaceResearch/TMD_Matlab_Toolbox_v2.5)) Matlab Toolbox was developed by Laurie Padman, Lana Erofeeva and Susan Howard. An updated version of the TMD Matlab Toolbox ([TMD3](https://github.com/chadagreene/Tide-Model-Driver)) was developed by Chad Greene. The OSU Tidal Inversion Software (OTIS) and OSU Tidal Prediction Software ([OTPS](https://www.tpxo.net/otps)) were developed by Lana Erofeeva and Gary Egbert ([copyright OSU](https://www.tpxo.net/tpxo-products-and-registration), licensed for non-commercial use). The NASA Goddard Space Flight Center (GSFC) PREdict Tidal Heights (PERTH3) software was developed by Richard Ray and Remko Scharroo. An updated and more versatile version of the NASA GSFC tidal prediction software ([PERTH5](https://codeberg.org/rray/perth5)) was developed by Richard Ray. ## License The content of this project is licensed under the [Creative Commons Attribution 4.0 Attribution license](https://creativecommons.org/licenses/by/4.0/) and the source code is licensed under the [MIT license](LICENSE).
text/markdown
Tyler Sutterley
tsutterl@uw.edu
pyTMD contributors
null
MIT License Copyright (c) 2017 Tyler C Sutterley Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
Ocean Tides, Load Tides, Pole Tides, Solid Earth Tides, Tidal Prediction
[ "Development Status :: 3 - Alpha", "Intended Audience :: Science/Research", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Topic :: Scientific/Engineering :: Physics", "Topic :: Scientific/Engineering :: Oceanography" ]
[]
null
null
~=3.9
[]
[]
[]
[ "h5netcdf", "lxml", "numpy", "pint", "platformdirs", "pyproj>=2.5.0", "scipy>=1.10.1", "timescale>=0.0.8", "xarray", "docutils>=0.17; extra == \"doc\"", "graphviz; extra == \"doc\"", "ipympl; extra == \"doc\"", "myst-nb; extra == \"doc\"", "numpydoc; extra == \"doc\"", "sphinx; extra == \"doc\"", "sphinx-argparse>=0.4; extra == \"doc\"", "sphinxcontrib-bibtex; extra == \"doc\"", "sphinx-design; extra == \"doc\"", "sphinx_rtd_theme; extra == \"doc\"", "cartopy; extra == \"all\"", "ipyleaflet; extra == \"all\"", "ipywidgets; extra == \"all\"", "jplephem; extra == \"all\"", "jupyterlab; extra == \"all\"", "matplotlib; extra == \"all\"", "notebook; extra == \"all\"", "pandas; extra == \"all\"", "dask; extra == \"aws\"", "obstore; extra == \"aws\"", "pandas; extra == \"aws\"", "pyarrow; extra == \"aws\"", "s3fs; extra == \"aws\"", "zarr>=3; extra == \"aws\"", "dask; extra == \"dev\"", "flake8; extra == \"dev\"", "gh; extra == \"dev\"", "oct2py; extra == \"dev\"", "pytest>=4.6; extra == \"dev\"", "pytest-cov; extra == \"dev\"", "pytest-xdist; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://pytmd.readthedocs.io", "Documentation, https://pytmd.readthedocs.io", "Repository, https://github.com/pyTMD/pyTMD", "Issues, https://github.com/pyTMD/pyTMD/issues" ]
twine/6.2.0 CPython/3.14.2
2026-02-21T01:04:15.902894
pytmd-3.0.3.tar.gz
383,625
e9/db/a14770912ce44d0a8cebd1f65916e9bfaa28082d6ad909ffdf2cb6158d74/pytmd-3.0.3.tar.gz
source
sdist
null
false
f445a978576ee9ecea1b4fed5e311d25
3ae5107c7f52f7c8af46c11a34663218b5463c470bc8ac5483be16abe2f16b64
e9dba14770912ce44d0a8cebd1f65916e9bfaa28082d6ad909ffdf2cb6158d74
null
[ "LICENSE" ]
0
2.4
hms-commander
0.2.1
HEC-HMS automation library for hydrologic modeling - Python API for project management, simulation execution, and results analysis
# hms-commander <p align="center"> <img src="hms-commander_logo.svg" width=70%> </p> [![PyPI version](https://badge.fury.io/py/hms-commander.svg)](https://pypi.org/project/hms-commander/) [![Documentation Status](https://readthedocs.org/projects/hms-commander/badge/?version=latest)](https://hms-commander.readthedocs.io/en/latest/?badge=latest) [![Python 3.10+](https://img.shields.io/badge/python-3.10+-blue.svg)](https://www.python.org/downloads/) [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/licenses/MIT) **[📖 Full Documentation](https://hms-commander.readthedocs.io/)** > **Beta Software - Engineering Oversight Required** > > This library is in active development and should be used with caution. Many workflows have only been tested with HEC-HMS example projects, not production watersheds. > > **Real-world hydrologic modeling requires professional engineering judgment.** Every watershed has unique characteristics and nuances that automated workflows cannot fully capture. AI agent workflows are tools to assist engineers, not replace them. > > **Human-in-the-Loop is essential.** Licensed Professional Engineers must pilot these systems, guide their application, and verify all outputs before use in engineering decisions. Always validate results against established engineering practices and local knowledge. ## Why HMS Commander? **HMS→RAS linked models are an industry standard** for watershed-to-river hydraulic analysis, yet there is no straightforward way to automate the linkage between HEC-HMS (hydrology) and HEC-RAS (hydraulics). This library exists to **bridge that gap**—extending the [ras-commander](https://github.com/gpt-cmdr/ras-commander) effort for HEC-RAS automation to include HEC-HMS workflows. While HEC-HMS provides robust internal functionality for standalone hydrologic models, the real power emerges when HMS hydrographs flow into RAS hydraulic models for flood inundation mapping, bridge analysis, and infrastructure design. **HMS Commander enables:** - Automated HMS simulation execution and results extraction - DSS file operations for seamless HMS→RAS boundary condition transfer - Consistent API patterns across both HMS and RAS automation - LLM-assisted workflows for complex multi-model scenarios **LLM Forward Hydrologic Modeling Automation** A Python library for automating HEC-HMS operations, built using [CLB Engineering's LLM Forward Approach](https://clbengineering.com/). Follows the architectural patterns established by [ras-commander](https://github.com/gpt-cmdr/ras-commander). ## LLM Forward Approach HMS Commander implements [CLB Engineering's five core principles](docs/CLB_ENGINEERING_APPROACH.md): 1. **GUI Verifiability** - All changes inspectable in HEC-HMS GUI (no coding required for QAQC) 2. **Traceability** - Complete audit trail of model modifications 3. **QAQC-able Workflows** - Automated quality checks with pass/fail criteria 4. **Non-Destructive Operations** - Original models preserved via cloning 5. **Professional Documentation** - Client-ready reports and modeling logs **Result:** Automate tedious tasks while maintaining professional engineering standards. ## ⚠️ Breaking Changes in v0.2.0 **Precipitation hyetograph methods now return DataFrame instead of ndarray** If upgrading from v0.1.x, note that `Atlas14Storm`, `FrequencyStorm`, and `ScsTypeStorm` now return `pd.DataFrame` with columns `['hour', 'incremental_depth', 'cumulative_depth']` instead of `np.ndarray`. **Quick Migration:** ```python # OLD (v0.1.x) hyeto = Atlas14Storm.generate_hyetograph(total_depth_inches=17.0, ...) total = hyeto.sum() peak = hyeto.max() # NEW (v0.2.0+) hyeto = Atlas14Storm.generate_hyetograph(total_depth_inches=17.0, ...) total = hyeto['cumulative_depth'].iloc[-1] peak = hyeto['incremental_depth'].max() ``` **Why this change?** Standardizes API for HMS→RAS integration and includes time axis. See [CHANGELOG.md](CHANGELOG.md) for complete migration guide. ## Features - **Project Management**: Initialize and manage HEC-HMS projects with DataFrames - **File Operations**: Read and modify basin, met, control, and gage files - **Simulation Execution**: Run HEC-HMS via Jython scripts (single, batch, parallel) - **Results Analysis**: Extract peak flows, volumes, hydrograph statistics - **DSS Integration**: Read/write DSS files (via ras-commander) - **GIS Extraction**: Export model elements to GeoJSON - **Clone Operations**: Non-destructive model cloning for QAQC workflows ## Installation ### From PyPI (Recommended) ```bash # Create conda environment (recommended) conda create -n hms python=3.11 conda activate hms # Install hms-commander pip install hms-commander # Verify installation python -c "import hms_commander; print(hms_commander.__version__)" ``` ### Optional Dependencies ```bash # DSS file support (requires Java 8+) pip install hms-commander[dss] # GIS features (geopandas, shapely) pip install hms-commander[gis] # All optional features pip install hms-commander[all] ``` ### From Source (Development) ```bash # Clone repository git clone https://github.com/gpt-cmdr/hms-commander.git cd hms-commander # Create development environment conda create -n hmscmdr_local python=3.11 conda activate hmscmdr_local # Install in editable mode with all dependencies pip install -e ".[all]" # Verify using local copy python -c "import hms_commander; print(hms_commander.__file__)" # Should show: /path/to/hms-commander/hms_commander/__init__.py ``` ## Quick Start ```python from hms_commander import ( init_hms_project, hms, HmsBasin, HmsControl, HmsCmdr, HmsResults ) # Initialize project init_hms_project( r"C:/HMS_Projects/MyProject", hms_exe_path=r"C:/HEC/HEC-HMS/4.9/hec-hms.cmd" ) # View project data print(hms.basin_df) print(hms.run_df) # Run simulation success = HmsCmdr.compute_run("Run 1") # Extract results peaks = HmsResults.get_peak_flows("results.dss") print(peaks) ``` ## Example Notebooks Comprehensive Jupyter notebooks demonstrating workflows: | Notebook | Description | |----------|-------------| | [01_multi_version_execution.ipynb](examples/01_multi_version_execution.ipynb) | Execute across multiple HMS versions | | [02_run_all_hms413_projects.ipynb](examples/02_run_all_hms413_projects.ipynb) | Batch processing of example projects | | [03_project_dataframes.ipynb](examples/03_project_dataframes.ipynb) | Explore project DataFrames and component structure | | [04_hms_workflow.ipynb](examples/04_hms_workflow.ipynb) | Complete HMS workflow from init to results | | [05_run_management.ipynb](examples/05_run_management.ipynb) | Comprehensive run configuration guide | | [clone_workflow.ipynb](examples/clone_workflow.ipynb) | Non-destructive QAQC with model cloning | **Run Configuration Management (Phase 1)**: ```python from hms_commander import HmsRun # Modify run parameters with validation HmsRun.set_description("Run 1", "Updated scenario", hms_object=hms) HmsRun.set_basin("Run 1", "Basin_Model", hms_object=hms) # Validates component exists! HmsRun.set_dss_file("Run 1", "output.dss", hms_object=hms) # Prevents HMS from auto-deleting runs with invalid component references ``` See [05_run_management.ipynb](examples/05_run_management.ipynb) for complete examples. ## Library Structure | Class | Purpose | |-------|---------| | `HmsPrj` | Project manager (stateful singleton) | | `HmsBasin` | Basin model operations (.basin) | | `HmsControl` | Control specifications (.control) | | `HmsMet` | Meteorologic models (.met) | | `HmsGage` | Time-series gages (.gage) | | `HmsRun` | Run configuration management (.run) **NEW Phase 1** | | `HmsCmdr` | Simulation execution engine | | `HmsJython` | Jython script generation | | `HmsDss` | DSS file operations | | `HmsResults` | Results extraction & analysis | | `HmsGeo` | GIS data extraction | | `HmsUtils` | Utility functions | ## Key Methods ### Project Management ```python init_hms_project(path, hms_exe_path) # Initialize project hms.basin_df # Basin models DataFrame hms.run_df # Simulation runs DataFrame ``` ### Basin Operations ```python HmsBasin.get_subbasins(basin_path) # Get all subbasins HmsBasin.get_loss_parameters(basin_path, subbasin) # Get loss params HmsBasin.set_loss_parameters(basin_path, subbasin, curve_number=80) ``` ### Run Configuration (NEW Phase 1) ```python HmsRun.set_description("Run 1", "Updated scenario", hms_object=hms) HmsRun.set_basin("Run 1", "Basin_Model", hms_object=hms) # Validates! HmsRun.set_precip("Run 1", "Met_Model", hms_object=hms) # Validates! HmsRun.set_control("Run 1", "Control_Spec", hms_object=hms) # Validates! HmsRun.set_dss_file("Run 1", "output.dss", hms_object=hms) ``` ### Simulation Execution ```python HmsCmdr.compute_run("Run 1") # Single run HmsCmdr.compute_parallel(["Run 1", "Run 2"], max_workers=2) # Parallel HmsCmdr.compute_batch(["Run 1", "Run 2", "Run 3"]) # Sequential ``` ### Results Analysis ```python HmsResults.get_peak_flows("results.dss") # Peak flow summary HmsResults.get_volume_summary("results.dss") # Runoff volumes HmsResults.get_hydrograph_statistics("results.dss", "Outlet") HmsResults.compare_runs(["run1.dss", "run2.dss"], "Outlet") ``` ### DSS Operations ```python HmsDss.get_catalog("results.dss") # List all paths HmsDss.read_timeseries("results.dss", pathname) # Read time series HmsDss.extract_hms_results("results.dss", result_type="flow") ``` ## Requirements - Python 3.10+ - pandas, numpy, tqdm, requests ### Optional - **DSS**: ras-commander, pyjnius (Java 8+) - **GIS**: geopandas, pyproj, shapely ## Related Projects - [ras-commander](https://github.com/billk-FM/ras-commander) - HEC-RAS automation - [HEC-HMS](https://www.hec.usace.army.mil/software/hec-hms/) - USACE software ## Author **William Katzenmeyer, PE, CFM** - [CLB Engineering](https://clbengineering.com/) ## License MIT License
text/markdown
null
"William Katzenmeyer, PE, CFM" <billk@clbengineering.com>
null
null
null
HEC-HMS, hydrology, hydrologic modeling, water resources, engineering, automation, simulation, DSS, GIS
[ "Development Status :: 3 - Alpha", "Intended Audience :: Science/Research", "Intended Audience :: Developers", "Operating System :: Microsoft :: Windows", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Topic :: Scientific/Engineering :: GIS", "Topic :: Scientific/Engineering :: Hydrology" ]
[]
null
null
>=3.10
[]
[]
[]
[ "pandas>=1.5.0", "numpy>=1.21.0", "pathlib", "tqdm", "requests", "pytest>=7.0; extra == \"dev\"", "black>=22.0; extra == \"dev\"", "flake8>=4.0; extra == \"dev\"", "jupyter; extra == \"dev\"", "geopandas>=0.12.0; extra == \"gis\"", "pyproj>=3.3.0; extra == \"gis\"", "shapely>=2.0.0; extra == \"gis\"", "pynhd>=0.19.0; extra == \"gis\"", "pygeohydro>=0.19.0; extra == \"gis\"", "ras-commander>=0.83.0; extra == \"dss\"", "pyjnius; extra == \"dss\"", "xarray>=2023.1.0; extra == \"aorc\"", "zarr>=2.13.0; extra == \"aorc\"", "s3fs>=2023.1.0; extra == \"aorc\"", "netCDF4>=1.6.0; extra == \"aorc\"", "rioxarray>=0.13.0; extra == \"aorc\"", "mkdocs>=1.5.0; extra == \"docs\"", "mkdocs-material>=9.0.0; extra == \"docs\"", "mkdocstrings[python]>=0.24.0; extra == \"docs\"", "mkdocs-jupyter>=0.24.0; extra == \"docs\"", "mkdocs-git-revision-date-localized-plugin>=1.2.0; extra == \"docs\"", "hms-commander[aorc,dev,docs,dss,gis]; extra == \"all\"" ]
[]
[]
[]
[ "Homepage, https://github.com/gpt-cmdr/hms-commander", "Documentation, https://gpt-cmdr.github.io/hms-commander/", "Repository, https://github.com/gpt-cmdr/hms-commander", "Issues, https://github.com/gpt-cmdr/hms-commander/issues" ]
twine/6.2.0 CPython/3.13.5
2026-02-21T01:03:25.526592
hms_commander-0.2.1.tar.gz
4,157,769
ba/d0/af5d1a00b6927f3a6ae9fb5f9454b33749419c1987da5c1d34dce34a431d/hms_commander-0.2.1.tar.gz
source
sdist
null
false
29158211a61f2846396cb8af5136f680
f197e587a4ca3749ef46f98a49ec6e9a4c0da7fa5f0ca4044d3cd1924cf849f6
bad0af5d1a00b6927f3a6ae9fb5f9454b33749419c1987da5c1d34dce34a431d
MIT
[ "LICENSE" ]
228
2.4
get-pfm
0.2.0
PFM - Pure Fucking Magic. AI agent output container format.
# .pfm — Pure Fucking Magic [![MIT License](https://img.shields.io/badge/license-MIT-blue.svg)](LICENSE) [![Python 3.11+](https://img.shields.io/badge/python-3.11%2B-blue.svg)](https://www.python.org/) [![Tests: 178 passing](https://img.shields.io/badge/tests-178%20passing-brightgreen.svg)]() [![npm: get-pfm](https://img.shields.io/badge/npm-get--pfm-cb3837.svg)](https://www.npmjs.com/package/get-pfm) [![Security: A+](https://img.shields.io/badge/security-A%2B-brightgreen.svg)]() [![Web Viewer](https://img.shields.io/badge/web-viewer%20%26%20converter-58a6ff.svg)](https://getpfm.io/) [![Chrome Extension](https://img.shields.io/badge/chrome-extension%20(in%20review)-4285F4.svg)](https://chromewebstore.google.com/) > A universal container format for AI agent output. The AI ecosystem has a format problem. Agents spit out markdown, JSON, YAML, plain text, structured logs — all different, all incompatible, all missing context. `.pfm` fixes that. One file. Every agent. All the context. ``` #!PFM/1.0 #@meta id: 7a3f9c21-... agent: claude-code model: claude-opus-4-6 created: 2026-02-16T20:02:37Z checksum: 049dcea0... #@index content 329 398 chain 735 330 tools 1073 188 #@content Your actual agent output goes here. Full multiline content, any format. #@chain User: Analyze this codebase Agent: I'll examine the structure... #@tools search(query="authentication") read_file("auth/handler.py") #!END ``` That's it. Human readable. Machine parseable. Indexed for speed. --- ## Why PFM Exists Every AI agent today outputs differently: - ChatGPT gives you markdown blobs - Claude gives you structured text - AutoGPT dumps JSON logs - LangChain has its own trace format - CrewAI, Autogen, Semantic Kernel — all different There's no standard way to capture **what an agent produced**, **how it got there**, **what tools it used**, and **whether the output is trustworthy**. If you want to share agent output, audit it, chain it, or verify it — you're on your own. `.pfm` is the container that wraps it all: | What | Where | |------|-------| | The actual output | `content` section | | The conversation that produced it | `chain` section | | Tool calls made | `tools` section | | Agent reasoning | `reasoning` section | | Performance data | `metrics` section | | Anything else | Custom sections | --- ## Installation ```bash pip install . # From source pip install get-pfm # From PyPI ``` ## Quick Start ### Python API ```python from pfm import PFMDocument, PFMReader # Create doc = PFMDocument.create(agent="my-agent", model="claude-opus-4-6") doc.add_section("content", "The analysis shows 3 critical findings...") doc.add_section("chain", "User: Analyze the repo\nAgent: Starting analysis...") doc.add_section("tools", "grep(pattern='TODO', path='src/')") doc.write("report.pfm") # Read (full parse) doc = PFMReader.read("report.pfm") print(doc.content) # "The analysis shows..." print(doc.agent) # "my-agent" print(doc.chain) # "User: Analyze the repo..." # Read (indexed — O(1) section access, only reads what you need) with PFMReader.open("report.pfm") as reader: content = reader.get_section("content") # Jumps directly by byte offset print(reader.meta["agent"]) print(reader.section_names) ``` ### CLI ```bash pip install get-pfm # Python npm install -g get-pfm # Node.js — both install the `pfm` command ``` ```bash # Create a .pfm file pfm create -a "my-agent" -m "gpt-4" -c "Hello world" -o output.pfm # Pipe from stdin echo "Hello" | pfm create -a cli -o hello.pfm # Inspect metadata and sections pfm inspect output.pfm # Read a specific section pfm read output.pfm content # Validate structure and checksum pfm validate output.pfm # Quick file identification pfm identify output.pfm # Convert formats pfm convert to json output.pfm -o output.json pfm convert to md output.pfm -o output.md pfm convert from json data.json -o imported.pfm pfm convert from csv data.csv -o imported.pfm # Export conversations to fine-tuning data pfm export ./conversations/ -o training.jsonl --format openai pfm export ./conversations/ -o training.jsonl --format alpaca pfm export ./conversations/ -o training.jsonl --format sharegpt pfm export chat.pfm -o training.jsonl # single file ``` ### Spells Every CLI command has a Harry Potter spell alias. Run `pfm spells` for the full spellbook. ```bash pfm accio report.pfm content # Summon a section pfm polyjuice report.pfm json # Transform to another format pfm fidelius report.pfm # Encrypt (Fidelius Charm) pfm revelio report.pfm.enc # Decrypt (Revelio) pfm unbreakable-vow report.pfm # Sign (Unbreakable Vow) pfm vow-kept report.pfm # Verify signature pfm prior-incantato report.pfm # Full provenance + integrity check pfm pensieve ./conversations/ # Extract training data (Pensieve) ``` ```python from pfm.spells import accio, polyjuice, fidelius, revelio, unbreakable_vow content = accio("report.pfm", "content") json_str = polyjuice(doc, "json") encrypted = fidelius(doc, "password") decrypted = revelio(encrypted, "password") unbreakable_vow(doc, "signing-key") ``` ### Chrome Extension Capture AI conversations directly from your browser. Available on the [Chrome Web Store](https://chromewebstore.google.com/) (in review). Supports ChatGPT, Claude, Gemini, Grok, and OpenClaw. One-click capture with optional encryption. **Features:** - Captures conversations from ChatGPT, Claude, and Gemini as .pfm files - Popup with Capture tab (save current conversation) and View/Convert tab (drop files) - Full-tab .pfm viewer with sidebar, search, keyboard shortcuts, and export - Zero dependencies — pure browser APIs only ### Converters Every format goes both ways: ```python from pfm import converters, PFMReader doc = PFMReader.read("report.pfm") # To other formats json_str = converters.to_json(doc) csv_str = converters.to_csv(doc) txt_str = converters.to_txt(doc) md_str = converters.to_markdown(doc) # From other formats doc = converters.from_json(json_str) doc = converters.from_csv(csv_str) doc = converters.from_txt("raw text", agent="importer") doc = converters.from_markdown(md_str) ``` ### Security ```python from pfm.security import sign, verify, encrypt_document, decrypt_document # Sign (HMAC-SHA256) doc = PFMDocument.create(agent="trusted-agent") doc.add_section("content", "verified output") sign(doc, secret="your-signing-key") doc.write("signed.pfm") # Verify loaded = PFMReader.read("signed.pfm") assert verify(loaded, "your-signing-key") # True if untampered # Encrypt (AES-256-GCM, PBKDF2 key derivation) encrypted = encrypt_document(doc, password="strong-password") with open("secret.pfm.enc", "wb") as f: f.write(encrypted) # Decrypt data = open("secret.pfm.enc", "rb").read() decrypted = decrypt_document(data, password="strong-password") print(decrypted.content) ``` --- ## Format Design Priorities In this exact order: ### 1. Speed - Magic byte check in first 64 bytes (instant file identification) - Index with byte offsets for O(1) section jumping (no scanning) - Two-pass writer pre-computes all offsets - Lazy reader only loads sections you request ### 2. Indexing - Every section has a byte offset and length in the `#@index` block - Readers can seek directly to any section without parsing the whole file - Multiple sections with the same name are supported and independently indexed ### 3. Human Readability - 100% UTF-8 text (no binary blobs) - Open it in any text editor and immediately understand the structure - Section markers (`#@`) are visually distinct and greppable - Magic line (`#!PFM/1.0`) is self-documenting ### 4. AI Usefulness - Structured metadata (agent, model, timestamps) - Prompt chain preservation (full conversation context) - Tool call logging (reproducibility) - Arbitrary custom sections (extend without breaking the spec) --- ## PFM vs Other Formats | Feature | .pfm | .json | .md | .yaml | .csv | |---------|-------|-------|-----|-------|------| | Human readable | Yes | Somewhat | Yes | Yes | Somewhat | | Indexed sections | **Yes (O(1))** | No | No | No | No | | Agent metadata | **Built-in** | Manual | No | Manual | No | | Prompt chain | **Built-in** | Manual | No | Manual | No | | Tool call logs | **Built-in** | Manual | No | Manual | No | | Checksum integrity | **Built-in** | No | No | No | No | | HMAC signing | **Built-in** | No | No | No | No | | Encryption | **AES-256-GCM** | No | No | No | No | | Multiline content | Natural | Escaped | Natural | Indented | Escaped | | File identification | 64 bytes | Parse whole file | No | Parse whole file | No | | Custom sections | Unlimited | N/A | N/A | N/A | N/A | | Bidirectional conversion | JSON, CSV, TXT, MD | — | — | — | — | ### Pros - **Purpose-built for AI** — not a general-purpose format retrofitted for agent output - **Self-contained** — one file has everything: output, context, metadata, provenance - **Fast** — indexed byte offsets mean you don't parse what you don't need - **Secure** — signing and encryption are first-class, not afterthoughts - **Extensible** — add any section you want, the format doesn't care - **Convertible** — round-trips cleanly to/from JSON, CSV, TXT, Markdown ### Cons - **New format** — no existing ecosystem (yet) - **Text-based** — not as compact as binary formats for large payloads - **Section markers in content** — content containing `#@` or `#!` on a line start requires `\` escaping (handled automatically by the library) - **Streaming write** — streaming mode uses trailing index, requires post-processing for full index - **Young spec** — v1.0, will evolve --- ## What It Solves 1. **Agent output portability** — Share agent results between tools, platforms, and teams in one format 2. **Provenance tracking** — Know which agent, model, and prompt produced any output 3. **Output verification** — Checksums and signatures prove content hasn't been tampered with 4. **Audit trails** — Chain and tool sections preserve the full generation context 5. **Selective reading** — Index-based access means you can grab just the section you need from large files 6. **Format interop** — Convert to/from JSON, CSV, TXT, MD without losing structure 7. **Training data export** — Export conversations to OpenAI, Alpaca, or ShareGPT JSONL for fine-tuning --- ## Project Structure ``` pfm/ ├── pyproject.toml # Package config, CLI entry point ├── pfm/ # Python library │ ├── spec.py # Format specification and constants │ ├── document.py # PFMDocument in-memory model │ ├── writer.py # Serializer with two-pass offset calculation │ ├── reader.py # Full parser + indexed lazy reader │ ├── stream.py # Streaming writer for real-time output │ ├── converters.py # JSON, CSV, TXT, Markdown (both directions) │ ├── export.py # Fine-tuning data export (OpenAI, Alpaca, ShareGPT) │ ├── security.py # HMAC signing, AES-256-GCM encryption │ ├── spells.py # Harry Potter spell aliases (accio, fidelius, etc.) │ ├── cli.py # Command-line interface (+ spell commands) │ ├── __main__.py # python -m pfm support │ ├── tui/ # Terminal viewer (Textual) │ └── web/ # Web viewer generator + local server ├── pfm-js/ # npm package (TypeScript) │ └── src/ # Parser, serializer, converters, checksum, CLI ├── pfm-vscode/ # VS Code extension │ └── src/ # Syntax, preview, outline, hover, CodeLens ├── pfm-chrome/ # Chrome extension (MV3) │ ├── content/ # AI site scrapers (ChatGPT, Claude, Gemini) │ ├── popup/ # Extension popup UI │ ├── viewer/ # Full-tab .pfm viewer │ └── shared/ # PFM core (parser, serializer, converters) ├── docs/ # GitHub Pages SPA (viewer & converter) ├── tests/ # 123 Python tests + conformance vectors └── examples/ └── hello.pfm # Example file ``` **123 Python tests. 55 JS tests. 178 total. All passing.** --- ## The Name Yes, PFM stands for **Pure Fucking Magic**. It's a 20-year joke. Someone in 2046 will google "what does .pfm stand for," find this page, and laugh. Then they'll realize the format actually works, and they'll keep using it. That's the plan. And honestly? The way AI agent data gets moved around today with zero standardization? The fact that it works at all *is* pure fucking magic. --- ## License MIT --- *Built in one session. Shipped before the hype cycle ended.*
text/markdown
Jason Sutter
null
null
null
null
ai, agent, file-format, container, pfm
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Software Development :: Libraries" ]
[]
null
null
>=3.11
[]
[]
[]
[ "pytest>=7.0; extra == \"dev\"", "pytest-cov>=4.0; extra == \"dev\"", "textual>=0.83.0; extra == \"tui\"" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.13.3
2026-02-21T01:00:29.597574
get_pfm-0.2.0.tar.gz
63,835
30/16/f0f6d22404ed84d6d28942e81cad1f00c1b6cf648fb42c3d7446cfaf41fa/get_pfm-0.2.0.tar.gz
source
sdist
null
false
54ff2f940dc25eb895a444f7b3851629
6e473b386ed6bfaf537e2351d406b190dc841908366b82bb11ef0919b1fa243d
3016f0f6d22404ed84d6d28942e81cad1f00c1b6cf648fb42c3d7446cfaf41fa
MIT
[ "LICENSE" ]
200
2.4
nano-crucible
2.0.0
Python Client for Crucible - National Archive for NSRC Observations
# nano-crucible : **N**ational **A**rchive for **N**SRC **O**bservations [![PyPI version](https://img.shields.io/pypi/v/nano-crucible.svg)](https://pypi.org/project/nano-crucible/) [![PyPI downloads](https://img.shields.io/pypi/dm/nano-crucible.svg)](https://pypi.org/project/nano-crucible/) [![GitHub release](https://img.shields.io/github/v/release/MolecularFoundryCrucible/nano-crucible)](https://github.com/MolecularFoundryCrucible/nano-crucible/releases) [![Python Version](https://img.shields.io/badge/python-3.8+-blue.svg)](https://www.python.org/downloads/) [![License](https://img.shields.io/badge/license-BSD-green.svg)](LICENSE) [![GitHub stars](https://img.shields.io/github/stars/MolecularFoundryCrucible/nano-crucible?style=social)](https://github.com/MolecularFoundryCrucible/nano-crucible) A Python client library and CLI tool for Crucible - the Molecular Foundry's data lakehouse for scientific research. Crucible stores experimental and synthetic data from DOE Nanoscale Science Research Centers (NSRCs), along with comprehensive metadata about samples, projects, instruments, and users. ## 🔬 What is Crucible? Crucible is the centralized data infrastructure for the [Molecular Foundry](https://foundry.lbl.gov/) and other [DOE Nanoscale Science Research Centers](https://science.osti.gov/bes/suf/User-Facilities/Nanoscale-Science-Research-Centers), providing: - **Unified data storage** for experimental and synthetic data - **Rich metadata** capture for to associate to datasets - **Sample provenance** tracking with parent-child relationships ## ✨ Features ### 🐍 Python API - **Dataset Management**: Create, query, update, and download datasets - **Sample Tracking**: Manage samples with hierarchical relationships and provenance - **Metadata**: Store and retrieve scientific metadata and experimental parameters - **Linking**: Connect datasets, samples, and create relationships programmatically ### 🖥️ Command-Line Interface - **`crucible config`**: One-time setup and configuration management - **`crucible upload`**: Upload datasets with automatic parsing and metadata extraction - **`crucible open`**: Open resources in the Crucible Web Explorer with one command - **`crucible link`**: Create relationships between datasets and samples ## 📦 Installation ### From PyPI (Recommended) ```bash pip install nano-crucible ``` ### From GitHub (Latest Development) ```bash pip install git+https://github.com/MolecularFoundryCrucible/nano-crucible ``` ### For Development ```bash git clone https://github.com/MolecularFoundryCrucible/nano-crucible.git cd nano-crucible pip install -e . ``` ## 🚀 Quick Start ### Python API #### Creating and Uploading Datasets ```python from crucible import CrucibleClient, BaseDataset from crucible.config import config # Get client client = config.client # Method 1: Create dataset (no files) dataset = client.create_new_dataset( unique_id = "my-unique-dataset-id", # Optional, auto-generated if None dataset_name="High-Temperature Synthesis", measurement="XRD", project_id="nanomaterials-2024", public=False, scientific_metadata={ "temperature_C": 800, "pressure_bar": 1.0, "duration_hours": 12, "atmosphere": "nitrogen" }, keywords=["synthesis", "high-temperature", "oxides"] ) # Method 2: Upload dataset with files using BaseDataset dataset = BaseDataset( unique_id="my-unique-dataset-id", # Optional, auto-generated if None dataset_name="Electron Microscopy Images", measurement="TEM", project_id="nanomaterials-2024", public=False, instrument_name="TEM-2100", data_format="TIFF", file_to_upload="/path/to/image.tiff" ) # Upload with metadata and files result = client.create_new_dataset_from_files( dataset=dataset, scientific_metadata={ "magnification": 50000, "voltage_kV": 200, "spot_size": 3 }, keywords=["TEM", "imaging", "nanoparticles"], files_to_upload=["/path/to/image.tiff", "/path/to/calibration.txt"], thumbnail="/path/to/thumbnail.png", # Optional ingestor='ApiUploadIngestor', wait_for_ingestion_response=True ) print(f"Dataset created: {result['created_record']['unique_id']}") ``` #### Linking Resources ```python # Link two datasets client.link_datasets("parent-dataset", "child-dataset") # Link two samples client.link_samples("parent-sample", "child-sample") # Link sample to dataset client.add_sample_to_dataset("dataset-id", "sample-id") ``` ### Command-Line Interface #### 1. Initial Configuration ```bash # One-time setup crucible config init # View your configuration crucible config show # Update settings crucible config set api_key YOUR_NEW_KEY ``` Get your API key at: [https://crucible.lbl.gov/api/v1/user_apikey](https://crucible.lbl.gov/api/v1/user_apikey) #### 2. Upload Data with Parsers ```bash # Upload with generic dataset crucible upload -i data.txt -pid my-project \ --metadata '{"temperature=300,pressure=1.0"}' \ --keywords "experiment,test" # Upload specific dataset (e.g. LAMMPS simulation) # Works only if the parser exists crucible upload -i simulation.lmp -t lammps -pid my-project ``` #### 3. Link Resources ```bash # Link two datasets crucible link -p parent_dataset_id -c child_dataset_id # Link two samples crucible link -p parent_sample_id -c child_sample_id # Link sample to dataset crucible link -d dataset_id -s sample_id ``` #### 4. Open in Browser ```bash # Open the Crucible Web Explorer crucible open # Open to a specific resource crucible open RESOURCE_MFID ``` ## 📖 Documentation - **CLI Documentation**: See [cli/README.md](crucible/cli/README.md) - **Parser Documentation**: See [parsers/README.md](crucible/parsers/README.md) - **API Reference**: Coming soon ## 🤝 Contributing We welcome contributions! Areas where you can help: - **New parsers** for additional data formats - **Bug reports** and feature requests - **Documentation** improvements - **Example notebooks** and tutorials ## 📄 License This project is licensed under the BSD-3-Clause License - see the [LICENSE](LICENSE) file for details. ## 🔗 Links - **Crucible API**: [https://crucible.lbl.gov/api/v1](https://crucible.lbl.gov/api/v1) - **Crucible Web Interface**: [https://crucible-graph-explorer.run.app](https://crucible-graph-explorer.run.app) ## 💬 Support For issues, questions, or feature requests: - **GitHub Issues**: [https://github.com/MolecularFoundryCrucible/nano-crucible/issues](https://github.com/MolecularFoundryCrucible/nano-crucible/issues) - **Email**: mkwall@lbl.gov, roncoroni@lbl.gov, esbarnard@lbl.gov --- **nano-crucible** is developed and maintained by the [Data Group](https://foundry.lbl.gov/expertise-instrumentation/#data-and-analytics-expertise) at the Molecular Foundry at Lawrence Berkeley National Laboratory.
text/markdown
null
mkywall <mkywall3@gmail.com>
null
null
BSD
crucible, nsrc, molecular-foundry, data-management, scientific-data, nanoscience
[ "Development Status :: 4 - Beta", "Intended Audience :: Science/Research", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Topic :: Scientific/Engineering", "Topic :: Scientific/Engineering :: Chemistry", "Topic :: Scientific/Engineering :: Physics" ]
[]
null
null
>=3.8
[]
[]
[]
[ "requests>=2.25.0", "pytz>=2021.1", "ipywidgets", "pydantic", "python-dotenv", "argcomplete>=2.0.0", "platformdirs>=2.5.0", "mfid>=1.0.0", "pytest>=6.0; extra == \"dev\"", "black>=21.0; extra == \"dev\"", "flake8>=3.8; extra == \"dev\"", "mypy>=0.812; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://github.com/MolecularFoundryCrucible/nano-crucible", "Repository, https://github.com/MolecularFoundryCrucible/nano-crucible", "Bug Tracker, https://github.com/MolecularFoundryCrucible/nano-crucible/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T01:00:04.966582
nano_crucible-2.0.0.tar.gz
39,746
70/76/6c07f1493123128e77d2f125fc966781beacdab07de5e0d79d223f01b873/nano_crucible-2.0.0.tar.gz
source
sdist
null
false
f96d6535e046060e91a36d532588dbb7
2cdd8a2872ecca13c9125445aa776180a3e82d23c79ba7ec4affe1239d95e3e2
70766c07f1493123128e77d2f125fc966781beacdab07de5e0d79d223f01b873
null
[ "LICENSE" ]
234
2.4
mockito
2.0.0a3
Spying framework
Mockito is a spying framework originally based on `the Java library with the same name <https://github.com/mockito/mockito>`_. (Actually *we* invented the strict stubbing mode back in 2009.) .. image:: https://github.com/kaste/mockito-python/actions/workflows/test-lint-go.yml/badge.svg :target: https://github.com/kaste/mockito-python/actions/workflows/test-lint-go.yml Install ======= ``pip install mockito`` Quick Start =========== 90% use case is that you want to stub out a side effect. :: from mockito import when, mock, unstub when(os.path).exists('/foo').thenReturn(True) # or: import requests # the famous library # you actually want to return a Response-like obj, we'll fake it response = mock({'status_code': 200, 'text': 'Ok'}) when(requests).get(...).thenReturn(response) # use it requests.get('http://google.com/') # clean up unstub() Read the docs ============= http://mockito-python.readthedocs.io/en/latest/ Breaking changes in v2 ====================== Two functions have been renamed: - `verifyNoMoreInteractions` is deprecated. Use `ensureNoUnverifiedInteractions` instead. Although `verifyNoMoreInteractions` is the name used in mockito for Java, it is a misnomer over there too (imo). Its docs say "Checks if any of given mocks has any unverified interaction.", and we make that clear now in the name of the function, so you don't need the docs to tell you what it does. - `verifyNoUnwantedInteractions` is deprecated. Use `verifyExpectedInteractions` instead. The new name should make it clear that it corresponds to the usage of `expect` (as alternative to `when`). Context managers now check the usage and any expectations (if you used `expect`) on exit. You can disable this check by setting the environment variable `MOCKITO_CONTEXT_MANAGERS_CHECK_USAGE` to `"0"`. Note that this does not disable the check for any explicit expectations you might have set with `expect`. This roughly corresponds to the `verifyStubbedInvocationsAreUsed` contra the `verifyExpectedInteractions` functions. New in v2 ========= - `between` now supports open ranges, e.g. `between=(0, )` to check that at least 0 interactions occurred. Development =========== I use `uv <https://docs.astral.sh/uv/>`_, and if you do too: you just clone this repo to your computer, then run ``uv sync`` in the root directory. Example usage:: uv run pytest Note: development and docs tooling target Python >=3.12, while the library itself supports older Python versions at runtime. For docs (Python >=3.12), install only the docs dependencies with:: uv sync --no-dev --group docs Or to install everything (all dependency groups, Python >=3.12), run:: uv sync --all-groups
text/x-rst
null
herr kaste <herr.kaste@gmail.com>
null
null
MIT
null
[]
[]
null
null
>=3.8
[]
[]
[]
[]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:57:56.592679
mockito-2.0.0a3.tar.gz
178,493
cb/2f/9de974e7ba86c38791340e0da9df01daa95fa2bf3749d1de0bef5b66bdbd/mockito-2.0.0a3.tar.gz
source
sdist
null
false
392a5738e2ed49e8a9b9e7f3fc7c5423
4bfb2c11b1a1fff0ebdb36b773117b70596bc36d41a93fcbc75908b94b9f0017
cb2f9de974e7ba86c38791340e0da9df01daa95fa2bf3749d1de0bef5b66bdbd
null
[ "AUTHORS", "LICENSE" ]
233
2.4
reaper-orchestra
0.1.1
Orchestra MCP server for REAPER
# Orchestra [理想中的音乐编辑模式](./docs/edit-music.md#设想中的最佳编辑方式) > 最好的 prompt 应该是 音乐概念 + 上下文 并重的 > 用户可以对指定的时间区域给出自己的指令,如: > “将选中的 vocal 区间加入 rubato 效果” > “为选中的 vocal 区间加入 guitar 配乐,同时需要参考此时其他配器轨道,注意音区避让” 但是目前的音乐生成模型还没有办法做到这一点。 比如 Suno,也只是给一个 style、lyrics,然后,开 roll,没有办法做到细粒度的编辑。 现在的模型缺乏音乐概念上的 prompt。 ## 软件架构 1. Reaper 宿主脚本模块(Lua 编写),通过文件系统与 Python MCP 通信 2. 各大音乐模型作为独立实体(Docker/Server/Process)暴露 MCP server,提供 `model.ability()` 作为调用 3. 添加 Skill,指引大模型(Claude/Gemini)进行工作流创作、辅助功能定义 ## 文档 1. [设想](./docs/edit-music.md) 2. [各大模型能力](./docs/music-model-abilities.md) 3. [开发任务追踪](./tasks/tasks.md) 4. [接入 MCP 规范](./docs/mcp/) Agent 代码阅读 resume checkpoint:019c79c7-2ff5-7273-8a64-c61bc751e965
text/markdown
null
null
null
null
null
null
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Operating System :: OS Independent" ]
[]
null
null
>=3.10
[]
[]
[]
[ "mcp>=1.25.0" ]
[]
[]
[]
[ "Homepage, https://github.com/rfhits/orchestra", "Repository, https://github.com/rfhits/orchestra" ]
twine/6.2.0 CPython/3.12.7
2026-02-21T00:57:33.923594
reaper_orchestra-0.1.1.tar.gz
569,163
e7/d5/3b286fd49396f19da8747a008f7a1948930924a3dbfb5ce162bb0c5c43b4/reaper_orchestra-0.1.1.tar.gz
source
sdist
null
false
f39f8b083fdd6e6b7056871ce2f1b32e
667cd576931234f9ceb6b450b6772da9b70367b2d277b2267c1584759db23778
e7d53b286fd49396f19da8747a008f7a1948930924a3dbfb5ce162bb0c5c43b4
null
[ "LICENSE" ]
223
2.4
fickling
0.1.8
A static analyzer and interpreter for Python pickle data
# Fickling ![Fickling image](./fickling_image.png) Fickling is a decompiler, static analyzer, and bytecode rewriter for Python [pickle](https://docs.python.org/3/library/pickle.html) object serializations. You can use fickling to detect, analyze, reverse engineer, or even create malicious pickle or pickle-based files, including PyTorch files. Fickling can be used both as a **python library** and a **CLI**. * [Installation](#installation) * [Securing AI/ML environments](#securing-aiml-environments) * [Generic malicious file detection](#generic-malicious-file-detection) * [Advanced usage](#advanced-usage) * [Trace pickle execution](#trace-pickle-execution) * [Pickle code injection](#pickle-code-injection) * [Pickle decompilation](#pickle-decompilation) * [PyTorch polyglots](#pytorch-polyglots) * [More information](#more-information) * [Contact](#contact) ## Installation Fickling has been tested on Python 3.9 through Python 3.13 and has very few dependencies. Both the library and command line utility can be installed through pip or uv: ```bash # Using pip python -m pip install fickling # Using uv uv pip install fickling ``` PyTorch is an optional dependency of Fickling. Therefore, in order to use Fickling's `pytorch` and `polyglot` modules, you should run: ```bash # Using pip python -m pip install fickling[torch] # Using uv uv pip install fickling[torch] ``` ## Securing AI/ML environments Fickling can help securing AI/ML codebases by automatically scanning pickle files contained in models. Fickling hooks the pickle module and verifies imports made when loading a model. It only checks the imports against an allowlist of imports from ML libraries that are considered safe, and blocks files that contain other imports. To enable Fickling security checks simply run the following lines once in your process, before loading any AI/ML models: ```python import fickling # This sets global hooks on pickle fickling.hook.activate_safe_ml_environment() ``` To remove the protection: ```python fickling.hook.deactivate_safe_ml_environment() ``` It is possible that the models you are using contain imports that aren't allowed by Fickling. If you still want to load the model, you can simply allow additional imports for your specific use-case with the `also_allow` argument: ```python fickling.hook.activate_safe_ml_environment(also_allow=[ "some.import", "another.allowed.import", ]) ``` **Important**: You should always make sure that manually added imports are actually safe and can not enable attackers to execute arbitrary code. If you are unsure on how to do that, you can open an issue on Fickling's Github repository that indicates the imports/models in question, and our team can review them and include them in the allow list if possible. ## Generic malicious file detection Fickling can seamlessly be integrated into your codebase to detect and halt the loading of malicious files at runtime. Below we show the different ways you can use fickling to enforce safety checks on pickle files. Under the hood, it hooks the `pickle` library to add safety checks so that loading a pickle file raises an `UnsafeFileError` exception if malicious content is detected in the file. #### Option 1 (recommended): check safety of all pickle files loaded ```python # This enforces safety checks every time pickle.load() is used fickling.always_check_safety() # Attempt to load an unsafe file now raises an exception with open("file.pkl", "rb") as f: try: pickle.load(f) except fickling.UnsafeFileError: print("Unsafe file!") ``` #### Option 2: use a context manager ```python with fickling.check_safety(): # All pickle files loaded within the context manager are checked for safety try: with open("file.pkl", "rb") as f: pickle.load("file.pkl") except fickling.UnsafeFileError: print("Unsafe file!") # Files loaded outside of context manager are NOT checked pickle.load("file.pkl") ``` #### Option 3: check and load a single file ```python # Use fickling.load() in place of pickle.load() to check safety and load a single pickle file try: fickling.load("file.pkl") except fickling.UnsafeFileError as e: print("Unsafe file!") ``` #### Option 4: only check pickle file safety without loading ```python3 # Perform a safety check on a pickle file without loading it if not fickling.is_likely_safe("file.pkl"): print("Unsafe file!") ``` #### Accessing the safety analysis results You can access the details of fickling's safety analysis from within the raised exception: ```python >>> try: ... fickling.load("unsafe.pkl") ... except fickling.UnsafeFileError as e: ... print(e.info) { "severity": "OVERTLY_MALICIOUS", "analysis": "Call to `eval(b'[5, 6, 7, 8]')` is almost certainly evidence of a malicious pickle file. Variable `_var0` is assigned value `eval(b'[5, 6, 7, 8]')` but unused afterward; this is suspicious and indicative of a malicious pickle file", "detailed_results": { "AnalysisResult": { "OvertlyBadEval": "eval(b'[5, 6, 7, 8]')", "UnusedVariables": [ "_var0", "eval(b'[5, 6, 7, 8]')" ] } } } ``` If you are using another language than Python, you can still use fickling's `CLI` to safety-check pickle files: ```console fickling --check-safety -p pickled.data ``` ## Advanced usage ### Trace pickle execution Fickling's `CLI` allows to safely trace the execution of the Pickle virtual machine without exercising any malicious code: ```console fickling --trace file.pkl ``` ### Pickle code injection Fickling allows to inject arbitrary code in a pickle file that will run every time the file is loaded ```console fickling --inject "print('Malicious')" file.pkl > malicious.pkl ``` ### Pickle decompilation Fickling can be used to decompile a pickle file for further analysis ```python >>> import ast, pickle >>> from fickling.fickle import Pickled >>> fickled_object = Pickled.load(pickle.dumps([1, 2, 3, 4])) >>> print(ast.dump(fickled_object.ast, indent=4)) Module( body=[ Assign( targets=[ Name(id='result', ctx=Store())], value=List( elts=[ Constant(value=1), Constant(value=2), Constant(value=3), Constant(value=4)], ctx=Load()))], type_ignores=[]) ``` ### PyTorch polyglots PyTorch contains multiple file formats with which one can make polyglot files, which are files that can be validly interpreted as more than one file format. Fickling supports identifying, inspecting, and creating polyglots with the following PyTorch file formats: * **PyTorch v0.1.1**: Tar file with sys_info, pickle, storages, and tensors * **PyTorch v0.1.10**: Stacked pickle files * **TorchScript v1.0**: ZIP file with model.json * **TorchScript v1.1**: ZIP file with model.json and attributes.pkl * **TorchScript v1.3**: ZIP file with data.pkl and constants.pkl * **TorchScript v1.4**: ZIP file with data.pkl, constants.pkl, and version set at 2 or higher (2 pickle files and a folder) * **PyTorch v1.3**: ZIP file containing data.pkl (1 pickle file) * **PyTorch model archive format[ZIP]**: ZIP file that includes Python code files and pickle files ```python >> import torch >> import torchvision.models as models >> from fickling.pytorch import PyTorchModelWrapper >> model = models.mobilenet_v2() >> torch.save(model, "mobilenet.pth") >> fickled_model = PyTorchModelWrapper("mobilenet.pth") >> print(fickled_model.formats) Your file is most likely of this format: PyTorch v1.3 ['PyTorch v1.3'] ``` Check out [our examples](https://github.com/trailofbits/fickling/tree/master/example) to learn more about using fickling! ## More information Pickled Python objects are in fact bytecode that is interpreted by a stack-based virtual machine built into Python called the "Pickle Machine". Fickling can take pickled data streams and decompile them into human-readable Python code that, when executed, will deserialize to the original serialized object. This is made possible by Fickling’s custom implementation of the PM. Fickling is safe to run on potentially malicious files because its PM symbolically executes code rather than overtly executing it. The authors do not prescribe any meaning to the “F” in Fickling; it could stand for “fickle,” … or something else. Divining its meaning is a personal journey in discretion and is left as an exercise to the reader. Learn more about fickling in our [blog post](https://blog.trailofbits.com/2021/03/15/never-a-dill-moment-exploiting-machine-learning-pickle-files/) and [DEF CON AI Village 2021 talk](https://www.youtube.com/watch?v=bZ0m_H_dEJI). ## Contact If you'd like to file a bug report or feature request, please use our [issues](https://github.com/trailofbits/fickling/issues) page. Feel free to contact us or reach out in [Empire Hacking](https://slack.empirehacking.nyc/) for help using or extending fickling. ## License This utility was developed by [Trail of Bits](https://www.trailofbits.com/). It is licensed under the [GNU Lesser General Public License v3.0](LICENSE). [Contact us](mailto:opensource@trailofbits.com) if you're looking for an exception to the terms. © 2021, Trail of Bits.
text/markdown
null
Trail of Bits <opensource@trailofbits.com>
null
Trail of Bits <opensource@trailofbits.com>
GNU LESSER GENERAL PUBLIC LICENSE Version 3, 29 June 2007 Copyright (C) 2007 Free Software Foundation, Inc. <https://fsf.org/> Everyone is permitted to copy and distribute verbatim copies of this license document, but changing it is not allowed. This version of the GNU Lesser General Public License incorporates the terms and conditions of version 3 of the GNU General Public License, supplemented by the additional permissions listed below. 0. Additional Definitions. As used herein, "this License" refers to version 3 of the GNU Lesser General Public License, and the "GNU GPL" refers to version 3 of the GNU General Public License. "The Library" refers to a covered work governed by this License, other than an Application or a Combined Work as defined below. An "Application" is any work that makes use of an interface provided by the Library, but which is not otherwise based on the Library. Defining a subclass of a class defined by the Library is deemed a mode of using an interface provided by the Library. A "Combined Work" is a work produced by combining or linking an Application with the Library. The particular version of the Library with which the Combined Work was made is also called the "Linked Version". The "Minimal Corresponding Source" for a Combined Work means the Corresponding Source for the Combined Work, excluding any source code for portions of the Combined Work that, considered in isolation, are based on the Application, and not on the Linked Version. The "Corresponding Application Code" for a Combined Work means the object code and/or source code for the Application, including any data and utility programs needed for reproducing the Combined Work from the Application, but excluding the System Libraries of the Combined Work. 1. Exception to Section 3 of the GNU GPL. You may convey a covered work under sections 3 and 4 of this License without being bound by section 3 of the GNU GPL. 2. Conveying Modified Versions. If you modify a copy of the Library, and, in your modifications, a facility refers to a function or data to be supplied by an Application that uses the facility (other than as an argument passed when the facility is invoked), then you may convey a copy of the modified version: a) under this License, provided that you make a good faith effort to ensure that, in the event an Application does not supply the function or data, the facility still operates, and performs whatever part of its purpose remains meaningful, or b) under the GNU GPL, with none of the additional permissions of this License applicable to that copy. 3. Object Code Incorporating Material from Library Header Files. The object code form of an Application may incorporate material from a header file that is part of the Library. You may convey such object code under terms of your choice, provided that, if the incorporated material is not limited to numerical parameters, data structure layouts and accessors, or small macros, inline functions and templates (ten or fewer lines in length), you do both of the following: a) Give prominent notice with each copy of the object code that the Library is used in it and that the Library and its use are covered by this License. b) Accompany the object code with a copy of the GNU GPL and this license document. 4. Combined Works. You may convey a Combined Work under terms of your choice that, taken together, effectively do not restrict modification of the portions of the Library contained in the Combined Work and reverse engineering for debugging such modifications, if you also do each of the following: a) Give prominent notice with each copy of the Combined Work that the Library is used in it and that the Library and its use are covered by this License. b) Accompany the Combined Work with a copy of the GNU GPL and this license document. c) For a Combined Work that displays copyright notices during execution, include the copyright notice for the Library among these notices, as well as a reference directing the user to the copies of the GNU GPL and this license document. d) Do one of the following: 0) Convey the Minimal Corresponding Source under the terms of this License, and the Corresponding Application Code in a form suitable for, and under terms that permit, the user to recombine or relink the Application with a modified version of the Linked Version to produce a modified Combined Work, in the manner specified by section 6 of the GNU GPL for conveying Corresponding Source. 1) Use a suitable shared library mechanism for linking with the Library. A suitable mechanism is one that (a) uses at run time a copy of the Library already present on the user's computer system, and (b) will operate properly with a modified version of the Library that is interface-compatible with the Linked Version. e) Provide Installation Information, but only if you would otherwise be required to provide such information under section 6 of the GNU GPL, and only to the extent that such information is necessary to install and execute a modified version of the Combined Work produced by recombining or relinking the Application with a modified version of the Linked Version. (If you use option 4d0, the Installation Information must accompany the Minimal Corresponding Source and Corresponding Application Code. If you use option 4d1, you must provide the Installation Information in the manner specified by section 6 of the GNU GPL for conveying Corresponding Source.) 5. Combined Libraries. You may place library facilities that are a work based on the Library side by side in a single library together with other library facilities that are not Applications and are not covered by this License, and convey such a combined library under terms of your choice, if you do both of the following: a) Accompany the combined library with a copy of the same work based on the Library, uncombined with any other library facilities, conveyed under the terms of this License. b) Give prominent notice with the combined library that part of it is a work based on the Library, and explaining where to find the accompanying uncombined form of the same work. 6. Revised Versions of the GNU Lesser General Public License. The Free Software Foundation may publish revised and/or new versions of the GNU Lesser General Public License from time to time. Such new versions will be similar in spirit to the present version, but may differ in detail to address new problems or concerns. Each version is given a distinguishing version number. If the Library as you received it specifies that a certain numbered version of the GNU Lesser General Public License "or any later version" applies to it, you have the option of following the terms and conditions either of that published version or of any later version published by the Free Software Foundation. If the Library as you received it does not specify a version number of the GNU Lesser General Public License, you may choose any version of the GNU Lesser General Public License ever published by the Free Software Foundation. If the Library as you received it specifies that a proxy can decide whether future versions of the GNU Lesser General Public License shall apply, that proxy's public statement of acceptance of any version is permanent authorization for you to choose that version for the Library.
null
[ "Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Science/Research", "License :: OSI Approved :: GNU Lesser General Public License v3 or later (LGPLv3+)", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Topic :: Security", "Topic :: Software Development :: Testing", "Topic :: Utilities" ]
[]
null
null
>=3.10
[]
[]
[]
[ "py7zr!=1.1.2,>=1.1.0; extra == \"archive\"", "coverage[toml]>=7.0.0; extra == \"dev\"", "numpy<2.3,>=2.2.6; python_version == \"3.10\" and extra == \"dev\"", "numpy>=2.3.5; python_version >= \"3.11\" and extra == \"dev\"", "py7zr!=1.1.2,>=1.1.0; extra == \"dev\"", "pytest-cov>=5.0.0; extra == \"dev\"", "pytest>=8.0.0; extra == \"dev\"", "ruff>=0.8.0; extra == \"dev\"", "torch>=2.1.0; extra == \"dev\"", "torchvision>=0.24.1; extra == \"dev\"", "ty>=0.0.12; extra == \"dev\"", "typing-extensions; python_version < \"3.12\" and extra == \"dev\"", "numpy; extra == \"examples\"", "pytorchfi; extra == \"examples\"", "ruff>=0.8.0; extra == \"lint\"", "ty>=0.0.12; extra == \"lint\"", "typing-extensions; python_version < \"3.12\" and extra == \"lint\"", "coverage[toml]>=7.0.0; extra == \"test\"", "numpy<2.3,>=2.2.6; python_version == \"3.10\" and extra == \"test\"", "numpy>=2.3.5; python_version >= \"3.11\" and extra == \"test\"", "py7zr!=1.1.2,>=1.1.0; extra == \"test\"", "pytest-cov>=5.0.0; extra == \"test\"", "pytest>=8.0.0; extra == \"test\"", "torch>=2.1.0; extra == \"test\"", "torchvision>=0.24.1; extra == \"test\"", "numpy<2.3,>=2.2.6; python_version == \"3.10\" and extra == \"torch\"", "numpy>=2.3.5; python_version >= \"3.11\" and extra == \"torch\"", "torch>=2.1.0; extra == \"torch\"", "torchvision>=0.24.1; extra == \"torch\"" ]
[]
[]
[]
[ "Homepage, https://pypi.org/project/fickling", "Issues, https://github.com/trailofbits/fickling/issues", "Source, https://github.com/trailofbits/fickling" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:57:26.106671
fickling-0.1.8.tar.gz
336,756
88/be/cd91e3921f064230ac9462479e4647fb91a7b0d01677103fce89f52e3042/fickling-0.1.8.tar.gz
source
sdist
null
false
4b4677e29243d9ba3cc51a9c3549e6fd
25a0bc7acda76176a9087b405b05f7f5021f76079aa26c6fe3270855ec57d9bf
88becd91e3921f064230ac9462479e4647fb91a7b0d01677103fce89f52e3042
null
[ "LICENSE" ]
3,168
2.3
agilicus
1.405.33
Agilicus SDK
## Agilicus SDK (Python) The [Agilicus Platform](https://www.agilicus.com/) [API](https://www.agilicus.com/api). is defined using [OpenAPI 3.0](https://github.com/OAI/OpenAPI-Specification), and may be used from any language. This allows configuration of our Zero-Trust Network Access cloud native platform using REST. You can see the API specification [online](https://www.agilicus.com/api). This package provides a Python SDK, class library interfaces for use in accessing individual collections. In addition it provides a command-line-interface (CLI) for interactive use. Read the class-library documentation [online](https://www.agilicus.com/api/) [Samples](https://git.agilicus.com/pub/samples) shows various examples of this code in use. Generally you may install this from [pypi](https://pypi.org/project/agilicus/) as: ``` pip install --upgrade agilicus ``` You may wish to add bash completion by adding this to your ~/.bashrc: ``` eval "$(_AGILICUS_CLI_COMPLETE=source agilicus-cli)" ``` ## Example: List users The below python code will show the same output as the CLI command: `agilicus-cli --issuer https://auth.dbt.agilicus.cloud list-users` ``` import agilicus import argparse import sys scopes = agilicus.scopes.DEFAULT_SCOPES parser = argparse.ArgumentParser(description="update-user") parser.add_argument("--auth-doc", type=str) parser.add_argument("--issuer", type=str) parser.add_argument("--email", type=str) parser.add_argument("--disable-user", type=bool, default=None) args = parser.parse_args() if not args.auth_doc and not args.issuer: print("error: specify either an --auth-doc or --issuer") sys.exit(1) if not args.email: print("error: specify an email to search for a user") sys.exit(1) api = agilicus.GetClient( agilicus_scopes=scopes, issuer=args.issuer, authentication_document=args.auth_doc ) users = api.users.list_users(org_id=api.default_org_id, email=args.email) if len(users.users) != 1: print(f"error: failed to find user with email: {args.email}") sys.exit(1) user = users.users[0] if args.disable_user is not None: user.enabled = args.disable_user result = api.users.replace_user( user.id, user=user, _check_input_type=False, _host_index=0 ) print(result) ```
text/markdown
Agilicus Devs
dev@agilicus.com
null
null
MIT
null
[ "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13" ]
[]
https://www.agilicus.com/
null
<4.0,>=3.10
[]
[]
[]
[ "certifi==2026.1.4", "python_dateutil>2.5.3", "PyJWT==2.11.0", "requests<3.0.0,>=2.23.0", "prettytable==3.17.0", "oauth2client<4.2.0,>=4.1.3", "click==8.3.1", "six<2.0.0,>=1.17.0", "cryptography>=39.0.0", "keyring>=23.0.0", "pyyaml==6.0.3", "stripe<15.0,>=2.60; extra == \"billing\"", "prometheus_client==0.24.1; extra == \"billing\"", "paho-mqtt<3.0.0,>=1.6.1", "colorama<0.5.0,>=0.4.6", "xdg<7.0.0,>=6.0.0", "keyrings.alt<6.0,>=4.2", "dateparser==1.3.0", "pem<24.0.0,>=23.1.0", "babel==2.18.0", "click-shell<3.0,>=2.1", "appdirs<2.0.0,>=1.4.4", "shortuuid<2.0.0,>=1.0.13", "httpx<0.29.0,>=0.28.1" ]
[]
[]
[]
[ "Homepage, https://www.agilicus.com/", "Repository, https://github.com/Agilicus" ]
poetry/2.1.4 CPython/3.12.12 Linux/6.17.8-061708-generic
2026-02-21T00:56:37.620072
agilicus-1.405.33.tar.gz
1,484,596
e9/76/361f16861261e4b83f145cd2dc15f0033e0f161b75ef1f36432dc3887484/agilicus-1.405.33.tar.gz
source
sdist
null
false
13d760b82e21e3529a1c21a01d7147ea
21126828c8f951bb1fb7475ab712686e5d81bc612302f5d491b2389b0dbe2a0f
e976361f16861261e4b83f145cd2dc15f0033e0f161b75ef1f36432dc3887484
null
[]
261
2.1
dda-py
0.4.1
Python bindings for DDA (Delay Differential Analysis)
# dda-py Python bindings for DDA (Delay Differential Analysis). The [DDA binary](https://snl.salk.edu/~sfdraeger/dda/) is required. Please download the most recent version from the file server. ## Installation ```bash pip install dda-py ``` Optional dependencies: ```bash pip install 'dda-py[mne]' # MNE-Python integration pip install 'dda-py[pandas]' # DataFrame export pip install 'dda-py[matplotlib]' # Plotting pip install 'dda-py[scipy]' # Window comparison statistics pip install 'dda-py[mne-bids]' # BIDS dataset integration pip install 'dda-py[all]' # All optional deps ``` ## Quick Start (High-Level API) ```python import numpy as np from dda_py import run_st # Analyze a numpy array (n_channels x n_samples) data = np.random.randn(3, 10000) result = run_st(data, sfreq=256.0, delays=(7, 10), wl=200, ws=100) print(result.coefficients.shape) # (3, n_windows, 3) print(result.n_channels) # 3 print(result.n_windows) # depends on data length print(result.to_dataframe().head()) ``` ### MNE-Python Integration ```python import mne from dda_py import run_st raw = mne.io.read_raw_edf("data.edf", preload=True) result = run_st(raw, delays=(7, 10), wl=200, ws=100) # sfreq is extracted automatically from the MNE Raw object ``` ### Cross-Timeseries Analysis ```python from dda_py import run_ct data = np.random.randn(4, 10000) # 4 channels result = run_ct(data, sfreq=256.0, delays=(7, 10), wl=200, ws=100) print(result.n_pairs) # 6 (all unique pairs) print(result.pair_labels) # ['ch0-ch1', 'ch0-ch2', ...] ``` ### Dynamical Ergodicity ```python from dda_py import run_de data = np.random.randn(2, 10000) result = run_de(data, sfreq=256.0, delays=(7, 10), wl=200, ws=100) print(result.ergodicity.shape) # (n_windows,) ``` ## Plotting Requires `pip install 'dda-py[matplotlib]'`. ```python from dda_py import run_st, plot_coefficients, plot_heatmap, plot_errors, plot_model result = run_st(data, sfreq=256.0, delays=(7, 10), wl=200, ws=100) # Coefficient time series per channel fig = plot_coefficients(result, use_time=True, sfreq=256.0) # Heatmap (channels x windows) for a single coefficient fig = plot_heatmap(result, coeff_index=0, cmap="RdBu_r") # Reconstruction errors over time fig = plot_errors(result) # Visualize the model space grid with selected terms highlighted fig = plot_model([1, 2, 10], num_delays=2, polynomial_order=4) ``` All plotting functions accept an optional `ax` parameter to draw into an existing matplotlib axes, and return a `matplotlib.figure.Figure`. ```python from dda_py import run_de, plot_ergodicity result = run_de(data, sfreq=256.0, delays=(7, 10), wl=200, ws=100) fig = plot_ergodicity(result, use_time=True, sfreq=256.0) ``` ## Batch Processing Process multiple files in one call: ```python from dda_py import run_batch, collect_results # Run DDA on a list of files results = run_batch( ["subj01.edf", "subj02.edf", "subj03.edf"], variant="st", sfreq=256.0, delays=(7, 10), wl=200, ws=100, progress=True, # shows progress bar (uses tqdm if installed) ) # Stack results into a single GroupResult for group analysis group = collect_results(results, labels=["subj01", "subj02", "subj03"]) print(group.coefficients.shape) # (3, n_channels, n_windows, n_coeffs) print(group.mean_over_windows()) # (3, n_channels, n_coeffs) print(group.to_dataframe().head()) ``` ## Statistics Group-level statistical analysis between two groups of DDA results. ### Permutation Test ```python from dda_py import permutation_test result = permutation_test( group_a=results_patients, group_b=results_controls, n_permutations=10000, seed=42, ) print(result.p_value) # (n_channels, n_coeffs) print(result.observed_stat) # (n_channels, n_coeffs) print(result.to_dataframe()) ``` ### Effect Size ```python from dda_py import compute_effect_size effect = compute_effect_size(results_patients, results_controls) print(effect.cohens_d) # (n_channels, n_coeffs) print(effect.to_dataframe()) ``` ### Window Comparison Requires `pip install 'dda-py[scipy]'`. Compare baseline vs test windows within a single recording: ```python from dda_py import compare_windows comp = compare_windows( result, baseline_windows=slice(0, 10), test_windows=slice(10, 20), method="ttest", # or "ranksum" ) print(comp.p_value) print(comp.baseline_mean) print(comp.test_mean) ``` ## BIDS Integration Requires `pip install 'dda-py[mne-bids]'`. Discover and analyze recordings from a BIDS dataset: ```python from dda_py import find_recordings, run_bids # List available recordings recordings = find_recordings("/path/to/bids", datatype="eeg", task="rest") for rec in recordings: print(rec.label) # e.g. "sub-01_ses-01_task-rest_run-01" # Run DDA on all matching recordings results = run_bids( "/path/to/bids", variant="st", datatype="eeg", task="rest", delays=(7, 10), wl=200, ws=100, ) # returns {"sub-01_task-rest": STResult, "sub-02_task-rest": STResult, ...} ``` ## Low-Level API For full control over the DDA binary: ```python from dda_py import DDARequest, DDARunner runner = DDARunner() # auto-discovers binary request = DDARequest( file_path="data.edf", channels=[0, 1, 2], variants=["ST"], window_length=200, window_step=100, delays=[7, 10], ) results = runner.run(request) ``` ## Model Encoding Visualize what DDA model indices mean: ```python from dda_py import visualize_model_space, decode_model_encoding # Show all monomials for 2 delays, polynomial order 4 print(visualize_model_space(2, 4, highlight_encoding=[1, 2, 10])) # Decode model [1, 2, 10] to equation print(decode_model_encoding([1, 2, 10], num_delays=2, polynomial_order=4, format="text")) # dx/dt = a_1 x_1 + a_2 x_2 + a_3 x_1^4 ``` ## CLI ```bash dda --file data.edf --channels 0 1 2 --variants ST --wl 200 --ws 100 dda --file data.edf --channels 0 1 2 --variants ST CT --delays 7 10 -o results.json ``` ## Variants - **ST** - Single Timeseries - **CT** - Cross Timeseries - **CD** - Cross Dynamical - **DE** - Dynamical Ergodicity - **SY** - Synchrony ## License MIT
text/markdown
null
Simon Draeger <sdraeger@salk.edu>
null
null
MIT
dda, delay-differential-analysis, eeg, neurophysiology, neuroscience, signal-processing
[ "Development Status :: 4 - Beta", "Intended Audience :: Science/Research", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Scientific/Engineering" ]
[]
null
null
>=3.9
[]
[]
[]
[ "numpy>=1.20", "matplotlib>=3.5; extra == \"all\"", "mne-bids>=0.12; extra == \"all\"", "mne>=1.0; extra == \"all\"", "pandas>=1.3; extra == \"all\"", "scipy>=1.7; extra == \"all\"", "matplotlib>=3.5; extra == \"dev\"", "mne>=1.0; extra == \"dev\"", "pandas>=1.3; extra == \"dev\"", "pytest-cov>=4.0; extra == \"dev\"", "pytest>=7.0; extra == \"dev\"", "scipy>=1.7; extra == \"dev\"", "matplotlib>=3.5; extra == \"matplotlib\"", "mne>=1.0; extra == \"mne\"", "mne-bids>=0.12; extra == \"mne-bids\"", "mne>=1.0; extra == \"mne-bids\"", "pandas>=1.3; extra == \"pandas\"", "scipy>=1.7; extra == \"scipy\"" ]
[]
[]
[]
[ "Homepage, https://github.com/sdraeger/dda-py", "Repository, https://github.com/sdraeger/dda-py", "Issues, https://github.com/sdraeger/dda-py/issues" ]
twine/6.2.0 CPython/3.14.3
2026-02-21T00:53:07.479772
dda_py-0.4.1.tar.gz
52,592
f4/79/eac056f22eb4f0453b3e0b00fe6c6483870b8f2176af3b789bb84be57ad6/dda_py-0.4.1.tar.gz
source
sdist
null
false
8ff90144153f85d960faad175b3bf007
276fe45b8c845d7772803f37ebe501fe315ad954f7ac4152a05b2bce5238df29
f479eac056f22eb4f0453b3e0b00fe6c6483870b8f2176af3b789bb84be57ad6
null
[]
235
2.4
Unified-python-sdk
0.57.8
Python Client SDK for Unified.to
<div align="left"> <a href="https://speakeasyapi.dev/"><img src="https://custom-icon-badges.demolab.com/badge/-Built%20By%20Speakeasy-212015?style=for-the-badge&logoColor=FBE331&logo=speakeasy&labelColor=545454" /></a> <a href="https://github.com/unified-to/unified-python-sdk/actions"><img src="https://img.shields.io/github/actions/workflow/status/unified-to/unified-python-sdk/speakeasy_sdk_generation.yml?style=for-the-badge" /></a> </div> <!-- Start Summary [summary] --> ## Summary Unified.to API: One API to Rule Them All For more information about the API: [API Documentation](https://docs.unified.to) <!-- End Summary [summary] --> <!-- Start Table of Contents [toc] --> ## Table of Contents <!-- $toc-max-depth=2 --> * [SDK Installation](https://github.com/unified-to/unified-python-sdk/blob/master/./#sdk-installation) * [IDE Support](https://github.com/unified-to/unified-python-sdk/blob/master/./#ide-support) * [SDK Example Usage](https://github.com/unified-to/unified-python-sdk/blob/master/./#sdk-example-usage) * [Available Resources and Operations](https://github.com/unified-to/unified-python-sdk/blob/master/./#available-resources-and-operations) * [File uploads](https://github.com/unified-to/unified-python-sdk/blob/master/./#file-uploads) * [Retries](https://github.com/unified-to/unified-python-sdk/blob/master/./#retries) * [Error Handling](https://github.com/unified-to/unified-python-sdk/blob/master/./#error-handling) * [Server Selection](https://github.com/unified-to/unified-python-sdk/blob/master/./#server-selection) * [Custom HTTP Client](https://github.com/unified-to/unified-python-sdk/blob/master/./#custom-http-client) * [Authentication](https://github.com/unified-to/unified-python-sdk/blob/master/./#authentication) * [Resource Management](https://github.com/unified-to/unified-python-sdk/blob/master/./#resource-management) * [Debugging](https://github.com/unified-to/unified-python-sdk/blob/master/./#debugging) <!-- End Table of Contents [toc] --> <!-- Start SDK Installation [installation] --> ## SDK Installation > [!NOTE] > **Python version upgrade policy** > > Once a Python version reaches its [official end of life date](https://devguide.python.org/versions/), a 3-month grace period is provided for users to upgrade. Following this grace period, the minimum python version supported in the SDK will be updated. The SDK can be installed with *uv*, *pip*, or *poetry* package managers. ### uv *uv* is a fast Python package installer and resolver, designed as a drop-in replacement for pip and pip-tools. It's recommended for its speed and modern Python tooling capabilities. ```bash uv add Unified-python-sdk ``` ### PIP *PIP* is the default package installer for Python, enabling easy installation and management of packages from PyPI via the command line. ```bash pip install Unified-python-sdk ``` ### Poetry *Poetry* is a modern tool that simplifies dependency management and package publishing by using a single `pyproject.toml` file to handle project metadata and dependencies. ```bash poetry add Unified-python-sdk ``` ### Shell and script usage with `uv` You can use this SDK in a Python shell with [uv](https://docs.astral.sh/uv/) and the `uvx` command that comes with it like so: ```shell uvx --from Unified-python-sdk python ``` It's also possible to write a standalone Python script without needing to set up a whole project like so: ```python #!/usr/bin/env -S uv run --script # /// script # requires-python = ">=3.10" # dependencies = [ # "Unified-python-sdk", # ] # /// from unified_python_sdk import UnifiedTo sdk = UnifiedTo( # SDK arguments ) # Rest of script here... ``` Once that is saved to a file, you can run it with `uv run script.py` where `script.py` can be replaced with the actual file name. <!-- End SDK Installation [installation] --> <!-- Start IDE Support [idesupport] --> ## IDE Support ### PyCharm Generally, the SDK will work well with most IDEs out of the box. However, when using PyCharm, you can enjoy much better integration with Pydantic by installing an additional plugin. - [PyCharm Pydantic Plugin](https://docs.pydantic.dev/latest/integrations/pycharm/) <!-- End IDE Support [idesupport] --> <!-- Start SDK Example Usage [usage] --> ## SDK Example Usage ### Example ```python # Synchronous Example from unified_python_sdk import UnifiedTo from unified_python_sdk.models import shared with UnifiedTo( security=shared.Security( jwt="<YOUR_API_KEY_HERE>", ), ) as unified_to: res = unified_to.accounting.create_accounting_account(request={ "accounting_account": {}, "connection_id": "<id>", }) assert res.accounting_account is not None # Handle response print(res.accounting_account) ``` </br> The same SDK client can also be used to make asynchronous requests by importing asyncio. ```python # Asynchronous Example import asyncio from unified_python_sdk import UnifiedTo from unified_python_sdk.models import shared async def main(): async with UnifiedTo( security=shared.Security( jwt="<YOUR_API_KEY_HERE>", ), ) as unified_to: res = await unified_to.accounting.create_accounting_account_async(request={ "accounting_account": {}, "connection_id": "<id>", }) assert res.accounting_account is not None # Handle response print(res.accounting_account) asyncio.run(main()) ``` <!-- End SDK Example Usage [usage] --> <!-- Start Available Resources and Operations [operations] --> ## Available Resources and Operations <details open> <summary>Available methods</summary> ### [Account](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/account/README.md) * [create_accounting_account](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/account/README.md#create_accounting_account) - Create an account * [get_accounting_account](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/account/README.md#get_accounting_account) - Retrieve an account * [list_accounting_accounts](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/account/README.md#list_accounting_accounts) - List all accounts * [patch_accounting_account](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/account/README.md#patch_accounting_account) - Update an account * [remove_accounting_account](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/account/README.md#remove_accounting_account) - Remove an account * [update_accounting_account](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/account/README.md#update_accounting_account) - Update an account ### [Accounting](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md) * [create_accounting_account](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#create_accounting_account) - Create an account * [create_accounting_bill](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#create_accounting_bill) - Create a bill * [create_accounting_category](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#create_accounting_category) - Create a category * [create_accounting_contact](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#create_accounting_contact) - Create a contact * [create_accounting_creditmemo](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#create_accounting_creditmemo) - Create a creditmemo * [create_accounting_expense](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#create_accounting_expense) - Create an expense * [create_accounting_invoice](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#create_accounting_invoice) - Create an invoice * [create_accounting_journal](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#create_accounting_journal) - Create a journal * [create_accounting_order](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#create_accounting_order) - Create an order * [create_accounting_purchaseorder](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#create_accounting_purchaseorder) - Create a purchaseorder * [create_accounting_salesorder](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#create_accounting_salesorder) - Create a salesorder * [create_accounting_taxrate](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#create_accounting_taxrate) - Create a taxrate * [create_accounting_transaction](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#create_accounting_transaction) - Create a transaction * [get_accounting_account](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#get_accounting_account) - Retrieve an account * [get_accounting_balancesheet](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#get_accounting_balancesheet) - Retrieve a balancesheet * [get_accounting_bill](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#get_accounting_bill) - Retrieve a bill * [get_accounting_cashflow](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#get_accounting_cashflow) - Retrieve a cashflow * [get_accounting_category](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#get_accounting_category) - Retrieve a category * [get_accounting_contact](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#get_accounting_contact) - Retrieve a contact * [get_accounting_creditmemo](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#get_accounting_creditmemo) - Retrieve a creditmemo * [get_accounting_expense](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#get_accounting_expense) - Retrieve an expense * [get_accounting_invoice](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#get_accounting_invoice) - Retrieve an invoice * [get_accounting_journal](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#get_accounting_journal) - Retrieve a journal * [get_accounting_order](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#get_accounting_order) - Retrieve an order * [get_accounting_organization](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#get_accounting_organization) - Retrieve an organization * [get_accounting_profitloss](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#get_accounting_profitloss) - Retrieve a profitloss * [get_accounting_purchaseorder](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#get_accounting_purchaseorder) - Retrieve a purchaseorder * [get_accounting_report](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#get_accounting_report) - Retrieve a report * [get_accounting_salesorder](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#get_accounting_salesorder) - Retrieve a salesorder * [get_accounting_taxrate](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#get_accounting_taxrate) - Retrieve a taxrate * [get_accounting_transaction](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#get_accounting_transaction) - Retrieve a transaction * [get_accounting_trialbalance](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#get_accounting_trialbalance) - Retrieve a trialbalance * [list_accounting_accounts](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#list_accounting_accounts) - List all accounts * [list_accounting_balancesheets](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#list_accounting_balancesheets) - List all balancesheets * [list_accounting_bills](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#list_accounting_bills) - List all bills * [list_accounting_cashflows](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#list_accounting_cashflows) - List all cashflows * [list_accounting_categories](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#list_accounting_categories) - List all categories * [list_accounting_contacts](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#list_accounting_contacts) - List all contacts * [list_accounting_creditmemoes](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#list_accounting_creditmemoes) - List all creditmemoes * [list_accounting_expenses](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#list_accounting_expenses) - List all expenses * [list_accounting_invoices](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#list_accounting_invoices) - List all invoices * [list_accounting_journals](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#list_accounting_journals) - List all journals * [list_accounting_orders](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#list_accounting_orders) - List all orders * [list_accounting_organizations](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#list_accounting_organizations) - List all organizations * [list_accounting_profitlosses](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#list_accounting_profitlosses) - List all profitlosses * [list_accounting_purchaseorders](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#list_accounting_purchaseorders) - List all purchaseorders * [list_accounting_reports](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#list_accounting_reports) - List all reports * [list_accounting_salesorders](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#list_accounting_salesorders) - List all salesorders * [list_accounting_taxrates](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#list_accounting_taxrates) - List all taxrates * [list_accounting_transactions](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#list_accounting_transactions) - List all transactions * [list_accounting_trialbalances](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#list_accounting_trialbalances) - List all trialbalances * [patch_accounting_account](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#patch_accounting_account) - Update an account * [patch_accounting_bill](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#patch_accounting_bill) - Update a bill * [patch_accounting_category](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#patch_accounting_category) - Update a category * [patch_accounting_contact](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#patch_accounting_contact) - Update a contact * [patch_accounting_creditmemo](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#patch_accounting_creditmemo) - Update a creditmemo * [patch_accounting_expense](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#patch_accounting_expense) - Update an expense * [patch_accounting_invoice](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#patch_accounting_invoice) - Update an invoice * [patch_accounting_journal](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#patch_accounting_journal) - Update a journal * [patch_accounting_order](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#patch_accounting_order) - Update an order * [patch_accounting_purchaseorder](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#patch_accounting_purchaseorder) - Update a purchaseorder * [patch_accounting_salesorder](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#patch_accounting_salesorder) - Update a salesorder * [patch_accounting_taxrate](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#patch_accounting_taxrate) - Update a taxrate * [patch_accounting_transaction](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#patch_accounting_transaction) - Update a transaction * [remove_accounting_account](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#remove_accounting_account) - Remove an account * [remove_accounting_bill](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#remove_accounting_bill) - Remove a bill * [remove_accounting_category](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#remove_accounting_category) - Remove a category * [remove_accounting_contact](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#remove_accounting_contact) - Remove a contact * [remove_accounting_creditmemo](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#remove_accounting_creditmemo) - Remove a creditmemo * [remove_accounting_expense](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#remove_accounting_expense) - Remove an expense * [remove_accounting_invoice](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#remove_accounting_invoice) - Remove an invoice * [remove_accounting_journal](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#remove_accounting_journal) - Remove a journal * [remove_accounting_order](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#remove_accounting_order) - Remove an order * [remove_accounting_purchaseorder](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#remove_accounting_purchaseorder) - Remove a purchaseorder * [remove_accounting_salesorder](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#remove_accounting_salesorder) - Remove a salesorder * [remove_accounting_taxrate](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#remove_accounting_taxrate) - Remove a taxrate * [remove_accounting_transaction](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#remove_accounting_transaction) - Remove a transaction * [update_accounting_account](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#update_accounting_account) - Update an account * [update_accounting_bill](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#update_accounting_bill) - Update a bill * [update_accounting_category](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#update_accounting_category) - Update a category * [update_accounting_contact](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#update_accounting_contact) - Update a contact * [update_accounting_creditmemo](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#update_accounting_creditmemo) - Update a creditmemo * [update_accounting_expense](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#update_accounting_expense) - Update an expense * [update_accounting_invoice](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#update_accounting_invoice) - Update an invoice * [update_accounting_journal](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#update_accounting_journal) - Update a journal * [update_accounting_order](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#update_accounting_order) - Update an order * [update_accounting_purchaseorder](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#update_accounting_purchaseorder) - Update a purchaseorder * [update_accounting_salesorder](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#update_accounting_salesorder) - Update a salesorder * [update_accounting_taxrate](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#update_accounting_taxrate) - Update a taxrate * [update_accounting_transaction](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/accounting/README.md#update_accounting_transaction) - Update a transaction ### [Activity](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/activity/README.md) * [create_ats_activity](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/activity/README.md#create_ats_activity) - Create an activity * [create_lms_activity](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/activity/README.md#create_lms_activity) - Create an activity * [get_ats_activity](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/activity/README.md#get_ats_activity) - Retrieve an activity * [get_lms_activity](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/activity/README.md#get_lms_activity) - Retrieve an activity * [list_ats_activities](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/activity/README.md#list_ats_activities) - List all activities * [list_lms_activities](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/activity/README.md#list_lms_activities) - List all activities * [patch_ats_activity](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/activity/README.md#patch_ats_activity) - Update an activity * [patch_lms_activity](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/activity/README.md#patch_lms_activity) - Update an activity * [remove_ats_activity](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/activity/README.md#remove_ats_activity) - Remove an activity * [remove_lms_activity](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/activity/README.md#remove_lms_activity) - Remove an activity * [update_ats_activity](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/activity/README.md#update_ats_activity) - Update an activity * [update_lms_activity](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/activity/README.md#update_lms_activity) - Update an activity ### [Ad](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ad/README.md) * [create_ads_ad](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ad/README.md#create_ads_ad) - Create an ad * [get_ads_ad](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ad/README.md#get_ads_ad) - Retrieve an ad * [list_ads_ads](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ad/README.md#list_ads_ads) - List all ads * [patch_ads_ad](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ad/README.md#patch_ads_ad) - Update an ad * [remove_ads_ad](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ad/README.md#remove_ads_ad) - Remove an ad * [update_ads_ad](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ad/README.md#update_ads_ad) - Update an ad ### [Ads](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md) * [create_ads_ad](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#create_ads_ad) - Create an ad * [create_ads_campaign](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#create_ads_campaign) - Create a campaign * [create_ads_creative](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#create_ads_creative) - Create a creative * [create_ads_group](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#create_ads_group) - Create a group * [create_ads_insertionorder](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#create_ads_insertionorder) - Create an insertionorder * [create_ads_organization](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#create_ads_organization) - Create an organization * [get_ads_ad](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#get_ads_ad) - Retrieve an ad * [get_ads_campaign](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#get_ads_campaign) - Retrieve a campaign * [get_ads_creative](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#get_ads_creative) - Retrieve a creative * [get_ads_group](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#get_ads_group) - Retrieve a group * [get_ads_insertionorder](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#get_ads_insertionorder) - Retrieve an insertionorder * [get_ads_organization](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#get_ads_organization) - Retrieve an organization * [list_ads_ads](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#list_ads_ads) - List all ads * [list_ads_campaigns](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#list_ads_campaigns) - List all campaigns * [list_ads_creatives](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#list_ads_creatives) - List all creatives * [list_ads_groups](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#list_ads_groups) - List all groups * [list_ads_insertionorders](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#list_ads_insertionorders) - List all insertionorders * [list_ads_organizations](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#list_ads_organizations) - List all organizations * [list_ads_reports](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#list_ads_reports) - List all reports * [patch_ads_ad](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#patch_ads_ad) - Update an ad * [patch_ads_campaign](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#patch_ads_campaign) - Update a campaign * [patch_ads_creative](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#patch_ads_creative) - Update a creative * [patch_ads_group](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#patch_ads_group) - Update a group * [patch_ads_insertionorder](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#patch_ads_insertionorder) - Update an insertionorder * [patch_ads_organization](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#patch_ads_organization) - Update an organization * [remove_ads_ad](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#remove_ads_ad) - Remove an ad * [remove_ads_campaign](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#remove_ads_campaign) - Remove a campaign * [remove_ads_creative](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#remove_ads_creative) - Remove a creative * [remove_ads_group](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#remove_ads_group) - Remove a group * [remove_ads_insertionorder](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#remove_ads_insertionorder) - Remove an insertionorder * [remove_ads_organization](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#remove_ads_organization) - Remove an organization * [update_ads_ad](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#update_ads_ad) - Update an ad * [update_ads_campaign](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#update_ads_campaign) - Update a campaign * [update_ads_creative](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#update_ads_creative) - Update a creative * [update_ads_group](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#update_ads_group) - Update a group * [update_ads_insertionorder](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#update_ads_insertionorder) - Update an insertionorder * [update_ads_organization](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ads/README.md#update_ads_organization) - Update an organization ### [Apicall](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/apicall/README.md) * [get_unified_apicall](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/apicall/README.md#get_unified_apicall) - Retrieve specific API Call by its ID * [list_unified_apicalls](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/apicall/README.md#list_unified_apicalls) - Returns API Calls ### [Application](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/application/README.md) * [create_ats_application](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/application/README.md#create_ats_application) - Create an application * [get_ats_application](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/application/README.md#get_ats_application) - Retrieve an application * [list_ats_applications](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/application/README.md#list_ats_applications) - List all applications * [patch_ats_application](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/application/README.md#patch_ats_application) - Update an application * [remove_ats_application](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/application/README.md#remove_ats_application) - Remove an application * [update_ats_application](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/application/README.md#update_ats_application) - Update an application ### [Applicationstatus](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/applicationstatus/README.md) * [list_ats_applicationstatuses](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/applicationstatus/README.md#list_ats_applicationstatuses) - List all applicationstatuses ### [Assessment](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/assessment/README.md) * [create_assessment_package](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/assessment/README.md#create_assessment_package) - Create an assessment package * [get_assessment_package](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/assessment/README.md#get_assessment_package) - Get an assessment package * [list_assessment_packages](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/assessment/README.md#list_assessment_packages) - List assessment packages * [patch_assessment_order](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/assessment/README.md#patch_assessment_order) - Update an order * [patch_assessment_package](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/assessment/README.md#patch_assessment_package) - Update an assessment package * [remove_assessment_package](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/assessment/README.md#remove_assessment_package) - Delete an assessment package * [update_assessment_order](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/assessment/README.md#update_assessment_order) - Update an order * [update_assessment_package](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/assessment/README.md#update_assessment_package) - Update an assessment package ### [Ats](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md) * [create_ats_activity](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#create_ats_activity) - Create an activity * [create_ats_application](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#create_ats_application) - Create an application * [create_ats_candidate](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#create_ats_candidate) - Create a candidate * [create_ats_company](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#create_ats_company) - Create a company * [create_ats_document](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#create_ats_document) - Create a document * [create_ats_interview](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#create_ats_interview) - Create an interview * [create_ats_job](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#create_ats_job) - Create a job * [create_ats_scorecard](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#create_ats_scorecard) - Create a scorecard * [get_ats_activity](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#get_ats_activity) - Retrieve an activity * [get_ats_application](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#get_ats_application) - Retrieve an application * [get_ats_candidate](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#get_ats_candidate) - Retrieve a candidate * [get_ats_company](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#get_ats_company) - Retrieve a company * [get_ats_document](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#get_ats_document) - Retrieve a document * [get_ats_interview](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#get_ats_interview) - Retrieve an interview * [get_ats_job](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#get_ats_job) - Retrieve a job * [get_ats_scorecard](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#get_ats_scorecard) - Retrieve a scorecard * [list_ats_activities](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#list_ats_activities) - List all activities * [list_ats_applications](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#list_ats_applications) - List all applications * [list_ats_applicationstatuses](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#list_ats_applicationstatuses) - List all applicationstatuses * [list_ats_candidates](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#list_ats_candidates) - List all candidates * [list_ats_companies](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#list_ats_companies) - List all companies * [list_ats_documents](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#list_ats_documents) - List all documents * [list_ats_interviews](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#list_ats_interviews) - List all interviews * [list_ats_jobs](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#list_ats_jobs) - List all jobs * [list_ats_scorecards](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#list_ats_scorecards) - List all scorecards * [patch_ats_activity](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#patch_ats_activity) - Update an activity * [patch_ats_application](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#patch_ats_application) - Update an application * [patch_ats_candidate](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#patch_ats_candidate) - Update a candidate * [patch_ats_company](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#patch_ats_company) - Update a company * [patch_ats_document](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#patch_ats_document) - Update a document * [patch_ats_interview](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#patch_ats_interview) - Update an interview * [patch_ats_job](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#patch_ats_job) - Update a job * [patch_ats_scorecard](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#patch_ats_scorecard) - Update a scorecard * [remove_ats_activity](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#remove_ats_activity) - Remove an activity * [remove_ats_application](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#remove_ats_application) - Remove an application * [remove_ats_candidate](https://github.com/unified-to/unified-python-sdk/blob/master/./docs/sdks/ats/README.md#remove
text/markdown
Unified API Inc.
null
null
null
null
null
[ "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14" ]
[]
https://github.com/unified-to/unified-python-sdk.git
null
>=3.10
[]
[]
[]
[ "httpcore>=1.0.9", "httpx>=0.28.1", "pydantic>=2.11.2" ]
[]
[]
[]
[ "Repository, https://github.com/unified-to/unified-python-sdk.git" ]
poetry/2.2.1 CPython/3.10.19 Linux/6.8.0-1044-azure
2026-02-21T00:52:18.129118
unified_python_sdk-0.57.8.tar.gz
569,739
59/5f/e4e473f0f3863780c6766018dd75349ee7a69fa51289c9c94b4ba19c4704/unified_python_sdk-0.57.8.tar.gz
source
sdist
null
false
f88417a4d4a19c993fc8e43d0b81a9dd
c16c491e4ac20383b244458e32e459b4fcaa196ed9562968154b6e82ef2d294b
595fe4e473f0f3863780c6766018dd75349ee7a69fa51289c9c94b4ba19c4704
null
[]
0
2.4
sf-config-builder
0.1.6
Manage Screaming Frog configs programmatically
# sf-config-builder Manage Screaming Frog `.seospiderconfig` files programmatically. ## Installation ```bash pip install sf-config-builder ``` ### Requirements - **Screaming Frog SEO Spider** must be installed (provides JARs for deserialization) - Python 3.8+ ## Quick Start ```python from sfconfig import SFConfig # Load existing config (auto-detects SF installation) config = SFConfig.load("base.seospiderconfig") # Or specify custom SF path config = SFConfig.load("base.seospiderconfig", sf_path="D:/Apps/Screaming Frog SEO Spider") # Configure for e-commerce audit config.max_urls = 100000 config.rendering_mode = "JAVASCRIPT" # Enable CSS/JS crawling (recommended for JS rendering) config.set("mCrawlConfig.mCrawlCSS", True) config.set("mCrawlConfig.mCrawlJavaScript", True) # Add custom extractions config.add_extraction("Price", "//span[@class='price']") config.add_extraction("SKU", "//span[@itemprop='sku']") config.add_extraction("Stock", ".availability", selector_type="CSS") # Add exclude patterns config.add_exclude(r".*\.pdf$") config.add_exclude(r".*/admin/.*") # Save and run config.save("client-audit.seospiderconfig") config.run_crawl("https://example.com", output_folder="./results") ``` ## Features ### Inspect Configs ```python config = SFConfig.load("my.seospiderconfig") # Get specific field max_urls = config.get("mCrawlConfig.mMaxUrls") # List all fields for field in config.fields(): print(f"{field['path']}: {field['value']}") # Filter by prefix crawl_fields = config.fields(prefix="mCrawlConfig") ``` ### Modify Configs ```python # Direct field access config.set("mCrawlConfig.mMaxUrls", 100000) # Convenience properties config.max_urls = 100000 config.max_depth = 10 config.rendering_mode = "JAVASCRIPT" # STATIC | JAVASCRIPT config.robots_mode = "IGNORE" # RESPECT | IGNORE config.crawl_delay = 0.5 config.user_agent = "MyBot/1.0" # CSS/JS crawling & storage config.set("mCrawlConfig.mCrawlCSS", True) config.set("mCrawlConfig.mStoreCSS", True) config.set("mCrawlConfig.mCrawlJavaScript", True) config.set("mCrawlConfig.mStoreJavaScript", True) # Performance / rate limiting config.set("mPerformanceConfig.mLimitPerformance", True) config.set("mPerformanceConfig.mUrlRequestsPerSecond", 5.0) ``` > **Note**: When switching to `JAVASCRIPT` rendering mode, Screaming Frog's GUI > automatically enables CSS/JS crawling. When building configs programmatically, > you should explicitly set `mCrawlCSS` and `mCrawlJavaScript` to `True` to > ensure proper rendering. ### Custom Extractions ```python # Add extraction rules config.add_extraction( name="Price", selector="//span[@class='price']", selector_type="XPATH", # XPATH | CSS | REGEX extract_mode="TEXT" # TEXT | HTML_ELEMENT | INNER_HTML ) # List extractions for ext in config.extractions: print(f"{ext['name']}: {ext['selector']}") # Remove by name config.remove_extraction("Price") # Clear all config.clear_extractions() ``` ### Custom Searches ```python # Add custom search filters config.add_custom_search( name="Filter 1", query=".*", mode="CONTAINS", data_type="REGEX", scope="HTML", case_sensitive=False ) # Remove by name config.remove_custom_search("Filter 1") # Clear all config.clear_custom_searches() ``` ### Custom JavaScript ```python # Add custom JavaScript extraction config.add_custom_javascript( name="Extractor 1", javascript="return document.title;", script_type="EXTRACTION", timeout_secs=10, content_types="text/html" ) # Remove by name config.remove_custom_javascript("Extractor 1") # Clear all config.clear_custom_javascript() ``` ### Exclude/Include Patterns ```python # Excludes (URLs matching these patterns are skipped) config.add_exclude(r".*\.pdf$") config.add_exclude(r".*/admin/.*") # Includes (only URLs matching these are crawled) config.add_include(r".*/products/.*") # List patterns print(config.excludes) print(config.includes) ``` ### Compare Configs ```python from sfconfig import SFConfig diff = SFConfig.diff("old.seospiderconfig", "new.seospiderconfig") if diff.has_changes: print(f"Found {diff.change_count} differences:") print(diff) # Filter by prefix crawl_changes = diff.changes_for("mCrawlConfig") ``` ### Test Extractions ```python # Test selector against live URL before full crawl result = config.test_extraction( url="https://example.com/product", selector="//span[@class='price']", selector_type="XPATH" ) if result["match_count"] > 0: print(f"Found: {result['matches']}") else: print("Selector didn't match - fix before crawling") ``` ### Run Crawls ```python # Blocking crawl config.run_crawl( url="https://example.com", output_folder="./results", export_tabs=["Internal:All", "Response Codes:All"], export_format="csv", timeout=3600 ) # Async crawl process = config.run_crawl_async( url="https://example.com", output_folder="./results" ) # Do other work... process.wait() # Block until complete ``` ## Multi-Client Workflow ```python from sfconfig import SFConfig clients = [ {"domain": "client1.com", "max_urls": 50000}, {"domain": "client2.com", "max_urls": 100000}, ] for client in clients: config = SFConfig.load("agency-base.seospiderconfig") config.max_urls = client["max_urls"] config.add_extraction("Price", "//span[@class='price']") config.save(f"/tmp/{client['domain']}.seospiderconfig") config.run_crawl( url=f"https://{client['domain']}", output_folder=f"./results/{client['domain']}" ) ``` ## Error Handling ```python from sfconfig import ( SFConfig, SFNotFoundError, SFValidationError, SFParseError, SFCrawlError ) try: config = SFConfig.load("my.seospiderconfig") config.set("mInvalidField", 123) config.save() except SFNotFoundError: print("Install Screaming Frog first") except SFValidationError as e: print(f"Invalid field: {e}") except SFParseError as e: print(f"Could not parse config: {e}") except SFCrawlError as e: print(f"Crawl failed: {e}") ``` ## Environment Variables | Variable | Description | |----------|-------------| | `SF_PATH` | Custom path to SF's JAR directory | | `SF_CLI_PATH` | Custom path to SF CLI executable | | `JAVA_HOME` | Custom Java installation path | ## Architecture ``` User Python code | v +------------------+ | sfconfig | (Python wrapper) | - SFConfig | | - SFDiff | +--------+---------+ | subprocess.run() v +------------------+ | ConfigBuilder | (Java CLI, bundled ~50KB) | .jar | +--------+---------+ | classpath includes v +------------------+ | SF's JARs | (from user's local SF install, NOT bundled) +------------------+ ``` At runtime, the library builds a classpath combining: - `ConfigBuilder.jar` (bundled with this package) - `{SF_INSTALL_PATH}/*` (user's local Screaming Frog JARs) This means: - Only our small JAR is distributed (no licensing issues) - SF's proprietary JARs are used from the user's existing installation - Compatibility is maintained across SF versions ## Development ### Building the Java CLI The Java source (`ConfigBuilder.java`) is included alongside the JAR in `sfconfig/java/`. To rebuild after modifications: ```bash cd sfconfig/java # Compile against SF's JARs (as compile-time dependency) javac -cp "/path/to/Screaming Frog SEO Spider/lib/*" ConfigBuilder.java # Package into JAR (include source for transparency) jar cfe ConfigBuilder.jar ConfigBuilder *.class ConfigBuilder.java # Clean up loose class files rm -f *.class ``` **Important**: Only bundle `ConfigBuilder.jar`. Do NOT bundle any JARs from SF's install directory - those are proprietary and already on the user's machine. ### Installing for Development ```bash cd sf-config-tool pip install -e ".[dev]" pytest tests/ ``` ## License MIT
text/markdown
Antonio
null
null
null
MIT
seo, screaming-frog, crawling, automation, config
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Topic :: Internet :: WWW/HTTP :: Indexing/Search", "Topic :: Software Development :: Libraries :: Python Modules" ]
[]
null
null
>=3.8
[]
[]
[]
[ "pytest>=7.0; extra == \"dev\"", "pytest-cov>=4.0; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://github.com/Amaculus/sf-config-builder", "Documentation, https://github.com/Amaculus/sf-config-builder#readme", "Repository, https://github.com/Amaculus/sf-config-builder", "Issues, https://github.com/Amaculus/sf-config-builder/issues" ]
twine/6.2.0 CPython/3.14.2
2026-02-21T00:51:18.213614
sf_config_builder-0.1.6.tar.gz
62,522
bf/ac/fa2d897dba6ff02e428ac46283dd92a19f487bb1061667ff598cdfd4c939/sf_config_builder-0.1.6.tar.gz
source
sdist
null
false
7fae4059fc27647c31583a672ce5c0f5
7cd38c1a2a73e7e8e7bd441eb2aba776de0f779c691778eba43111e2b2390fa8
bfacfa2d897dba6ff02e428ac46283dd92a19f487bb1061667ff598cdfd4c939
null
[ "LICENSE" ]
222
2.4
robo_appian
0.0.39
Automate your Appian code testing with Python. Boost quality, save time.
# Robo Appian Robo Appian is a Python library for automated UI testing of Appian applications. It provides user-friendly utilities and best practices to help you write robust, maintainable, and business-focused test automation. ## Features - Simple, readable API for Appian UI automation - Utilities for buttons, inputs, dropdowns, tables, tabs, and more - Data-driven and workflow testing support - Error handling and debugging helpers - Designed for both technical and business users ## Documentation Full documentation, guides, and API reference are available at: ➡️ [Robo Appian Documentation](https://dinilmithra.github.io/robo_appian/) ## Quick Start 1. Install Robo Appian: ```bash pip install robo_appian ``` 2. See the [Getting Started Guide](docs/getting-started/installation.md) for setup and your first test. ## Example Usage ```python from robo_appian.components import InputUtils, ButtonUtils # Set value in a text field by label InputUtils.setValueByLabelText(wait, "Username", "testuser") # Click a button by label ButtonUtils.clickByLabelText(wait, "Sign In") ``` ## Project Structure - `robo_appian/` - Library source code - `docs/` - Documentation and guides ## Contributing Contributions are welcome! Please see the [contributing guidelines](CONTRIBUTING.md) or open an issue to get started. ## License MIT License. See [LICENSE](LICENSE) for details. --- For questions or support, contact [Dinil Mithra](mailto:dinilmithra.mailme@gmail.com) or connect on [LinkedIn](https://www.linkedin.com/in/dinilmithra).
text/markdown
Dinil Mithra
dinilmithra.mailme@gmail.com
null
null
Apache-2.0
null
[ "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14" ]
[]
null
null
<4.0,>=3.12
[]
[]
[]
[ "numpy", "requests<3.0.0,>=2.25.1", "selenium>=4.34.0" ]
[]
[]
[]
[ "Homepage, https://github.com/dinilmithra/robo_appian", "Repository, https://github.com/dinilmithra/robo_appian.git" ]
poetry/2.3.2 CPython/3.12.1 Linux/6.8.0-1030-azure
2026-02-21T00:50:47.100225
robo_appian-0.0.39-py3-none-any.whl
31,959
1b/7e/1e404cc6d3a8b14d267d275d2cfde8435a82000533834ad084e43212aea5/robo_appian-0.0.39-py3-none-any.whl
py3
bdist_wheel
null
false
e58cadc013a52e5b8c3225b8b22963df
f102b539ab37489e1ec5cc788012d3af1e6794984b80953d8ee46ac1eeb8458e
1b7e1e404cc6d3a8b14d267d275d2cfde8435a82000533834ad084e43212aea5
null
[ "LICENSE" ]
0
2.4
Geode-Implicit
4.5.4
Licensed framework for working with implicit modeling
<h1 align="center">Geode-Implicit<sup><i>by Geode-solutions</i></sup></h1> <h3 align="center">Module for working with implicit modeling.</h3> <p align="center"> <img src="https://github.com/Geode-solutions/OpenGeode-ModuleTemplate/workflows/CI/badge.svg" alt="Build Status"> <img src="https://github.com/Geode-solutions/OpenGeode-ModuleTemplate/workflows/CD/badge.svg" alt="Deploy Status"> <img src="https://codecov.io/gh/Geode-solutions/OpenGeode-ModuleTemplate/branch/master/graph/badge.svg" alt="Coverage Status"> <img src="https://img.shields.io/github/release/Geode-solutions/OpenGeode-ModuleTemplate.svg" alt="Version"> </p> <p align="center"> <img src="https://img.shields.io/static/v1?label=Windows&logo=windows&logoColor=white&message=support&color=success" alt="Windows support"> <img src="https://img.shields.io/static/v1?label=Ubuntu&logo=Ubuntu&logoColor=white&message=support&color=success" alt="Ubuntu support"> <img src="https://img.shields.io/static/v1?label=Red%20Hat&logo=Red-Hat&logoColor=white&message=support&color=success" alt="Red Hat support"> <img src="https://img.shields.io/static/v1?label=macOS&logo=apple&logoColor=white&message=support&color=success" alt="macOS support"> </p> <p align="center"> <img src="https://img.shields.io/badge/C%2B%2B-11-blue.svg" alt="Language"> <img src="https://img.shields.io/badge/license-MIT-blue.svg" alt="License"> <img src="https://img.shields.io/badge/%20%20%F0%9F%93%A6%F0%9F%9A%80-semantic--release-e10079.svg" alt="Semantic-release"> <a href="https://geode-solutions.com/#slack"> <img src="https://opengeode-slack-invite.herokuapp.com/badge.svg" alt="Slack invite"> </a> <a href="https://doi.org/10.5281/zenodo.3610370"> <img src="https://zenodo.org/badge/DOI/10.5281/zenodo.3610370.svg" alt="DOI"> </a> --- ## Introduction Geode-Implicit is an [OpenGeode] module for working with implicit modeling, from the construction of an implicit model using data points to its explicitation in a topologically valid meshed model. [OpenGeode]: https://github.com/Geode-solutions/OpenGeode Copyright (c) 2019 - 2026, Geode-solutions
text/markdown
null
Geode-solutions <contact@geode-solutions.com>
null
null
Proprietary
null
[]
[]
null
null
<3.13,>=3.9
[]
[]
[]
[ "geode-background==9.*,>=9.10.3", "geode-common==33.*,>=33.19.0", "geode-conversion==6.*,>=6.5.14", "geode-explicit==6.*,>=6.7.4", "geode-numerics==6.*,>=6.4.12", "geode-simplex==11.*,>=11.0.0", "opengeode-core==15.*,>=15.31.5", "opengeode-geosciences==9.*,>=9.5.9", "opengeode-geosciencesio==5.*,>=5.8.10", "opengeode-inspector==6.*,>=6.8.17", "opengeode-io==7.*,>=7.4.8" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.9.13
2026-02-21T00:49:26.169262
geode_implicit-4.5.4-cp39-cp39-win_amd64.whl
7,341,672
c6/58/fa1e1ce9ab55a5fed083f25ce3c2b7b1f3b997464054f536f6c46e2bf39f/geode_implicit-4.5.4-cp39-cp39-win_amd64.whl
cp39
bdist_wheel
null
false
110943b853efe93afe4f93ba8ccf413a
fea4c5c1efb773046efb95336e7d4f53f86f6d7631edec147127048e9209b7f1
c658fa1e1ce9ab55a5fed083f25ce3c2b7b1f3b997464054f536f6c46e2bf39f
null
[]
0
2.4
fastbinning
0.0.1
A high-performance binning library specifically designed for Credit Risk Modeling and Scorecard Development.
<div style="text-align: center;"> <img src="https://capsule-render.vercel.app/api?type=transparent&height=300&color=gradient&text=fastbinning&section=header&reversal=false&height=120&fontSize=90"> </div> <p align="center"> <a href="https://github.com/RektPunk/fastbinning/releases/latest"> <img alt="release" src="https://img.shields.io/github/v/release/RektPunk/fastbinning.svg"> </a> <a href="https://github.com/RektPunk/fastbinning/blob/main/LICENSE"> <img alt="License" src="https://img.shields.io/github/license/RektPunk/fastbinning.svg"> </a> </p> A high-performance binning library specifically designed for **Credit Risk Modeling** and **Scorecard Development**. In financial risk modeling, **Weight of Evidence (WoE)** and **Information Value (IV)** are gold standards for feature engineering. `fastbinning` ensures mathematical rigor with extreme speed. # Why fastbinning for Credit Scoring? * **Monotonicity Guaranteed**: In credit scoring, features like 'Utilization Rate' or 'Age' must have a monotonic relationship with default risk to be explainable and compliant. * **Built for Big Data**: While traditional tools struggle with millions of rows, `fastbinning` handles 10M+ records in milliseconds. * **Robustness**: Prevents overfitting by enforcing minimum sample constraints (`min_bin_pct`), ensuring each bin is statistically significant. # Installation Install using pip: ```bash pip install fastbinning ``` # Example Please refer to the [**Examples**](https://github.com/RektPunk/fastbinning/tree/main/examples) provided for further clarification.
text/markdown; charset=UTF-8; variant=GFM
null
RektPunk <rektpunk@gmail.com>
null
null
null
null
[]
[]
null
null
>=3.8
[]
[]
[]
[ "numpy>=1.21.6" ]
[]
[]
[]
[ "repository, https://github.com/RektPunk/rektrag" ]
maturin/1.12.3
2026-02-21T00:49:24.739709
fastbinning-0.0.1.tar.gz
79,319
ef/12/ab4a829294ce8ecc4cc205a2a6150b5a6570a23e641c4a74ab7a966e537d/fastbinning-0.0.1.tar.gz
source
sdist
null
false
85390394f6aac66ef5acc3171e9a0025
fbf331d1e7fedf37d19c4f8ec4d42f85b377440a462336d35a5384c286d145d2
ef12ab4a829294ce8ecc4cc205a2a6150b5a6570a23e641c4a74ab7a966e537d
null
[]
327
2.4
scrapy-impersonate
1.6.3
Scrapy download handler that can impersonate browser fingerprints
# scrapy-impersonate [![version](https://img.shields.io/pypi/v/scrapy-impersonate.svg)](https://pypi.python.org/pypi/scrapy-impersonate) `scrapy-impersonate` is a Scrapy download handler. This project integrates [curl_cffi](https://github.com/yifeikong/curl_cffi) to perform HTTP requests, so it can impersonate browsers' TLS signatures or JA3 fingerprints. ## Installation ``` pip install scrapy-impersonate ``` ## Activation To use this package, replace the default `http` and `https` Download Handlers by updating the [`DOWNLOAD_HANDLERS`](https://docs.scrapy.org/en/latest/topics/settings.html#download-handlers) setting: ```python DOWNLOAD_HANDLERS = { "http": "scrapy_impersonate.ImpersonateDownloadHandler", "https": "scrapy_impersonate.ImpersonateDownloadHandler", } ``` By setting `USER_AGENT = None`, `curl_cffi` will automatically choose the appropriate User-Agent based on the impersonated browser: ```python USER_AGENT = "" ``` Also, be sure to [install the asyncio-based Twisted reactor](https://docs.scrapy.org/en/latest/topics/asyncio.html#installing-the-asyncio-reactor) for proper asynchronous execution: ```python TWISTED_REACTOR = "twisted.internet.asyncioreactor.AsyncioSelectorReactor" ``` ## Usage Set the `impersonate` [Request.meta](https://docs.scrapy.org/en/latest/topics/request-response.html#scrapy.http.Request.meta) key to download a request using `curl_cffi`: ```python import scrapy class ImpersonateSpider(scrapy.Spider): name = "impersonate_spider" custom_settings = { "TWISTED_REACTOR": "twisted.internet.asyncioreactor.AsyncioSelectorReactor", "USER_AGENT": "", "DOWNLOAD_HANDLERS": { "http": "scrapy_impersonate.ImpersonateDownloadHandler", "https": "scrapy_impersonate.ImpersonateDownloadHandler", }, "DOWNLOADER_MIDDLEWARES": { "scrapy_impersonate.RandomBrowserMiddleware": 1000, }, } def start_requests(self): for _ in range(5): yield scrapy.Request( "https://tls.browserleaks.com/json", dont_filter=True, ) def parse(self, response): # ja3_hash: 98cc085d47985d3cca9ec1415bbbf0d1 (chrome133a) # ja3_hash: 2d692a4485ca2f5f2b10ecb2d2909ad3 (firefox133) # ja3_hash: c11ab92a9db8107e2a0b0486f35b80b9 (chrome124) # ja3_hash: 773906b0efdefa24a7f2b8eb6985bf37 (safari15_5) # ja3_hash: cd08e31494f9531f560d64c695473da9 (chrome99_android) yield {"ja3_hash": response.json()["ja3_hash"]} ``` ### impersonate-args You can pass any necessary [arguments](https://github.com/lexiforest/curl_cffi/blob/38a91f2e7b23d9c9bda1d8085b7e41e33767c768/curl_cffi/requests/session.py#L1189-L1222) to `curl_cffi` through `impersonate_args`. For example: ```python yield scrapy.Request( "https://tls.browserleaks.com/json", dont_filter=True, meta={ "impersonate": browser, "impersonate_args": { "verify": False, "timeout": 10, }, }, ) ``` ## Supported browsers The following browsers can be impersonated | Browser | Version | Build | OS | Name | | --- | --- | --- | --- | --- | | ![Chrome](https://raw.githubusercontent.com/alrra/browser-logos/main/src/chrome/chrome_24x24.png "Chrome") | 99 | 99.0.4844.51 | Windows 10 | `chrome99` | | ![Chrome](https://raw.githubusercontent.com/alrra/browser-logos/main/src/chrome/chrome_24x24.png "Chrome") | 99 | 99.0.4844.73 | Android 12 | `chrome99_android` | | ![Chrome](https://raw.githubusercontent.com/alrra/browser-logos/main/src/chrome/chrome_24x24.png "Chrome") | 100 | 100.0.4896.75 | Windows 10 | `chrome100` | | ![Chrome](https://raw.githubusercontent.com/alrra/browser-logos/main/src/chrome/chrome_24x24.png "Chrome") | 101 | 101.0.4951.67 | Windows 10 | `chrome101` | | ![Chrome](https://raw.githubusercontent.com/alrra/browser-logos/main/src/chrome/chrome_24x24.png "Chrome") | 104 | 104.0.5112.81 | Windows 10 | `chrome104` | | ![Chrome](https://raw.githubusercontent.com/alrra/browser-logos/main/src/chrome/chrome_24x24.png "Chrome") | 107 | 107.0.5304.107 | Windows 10 | `chrome107` | | ![Chrome](https://raw.githubusercontent.com/alrra/browser-logos/main/src/chrome/chrome_24x24.png "Chrome") | 110 | 110.0.5481.177 | Windows 10 | `chrome110` | | ![Chrome](https://raw.githubusercontent.com/alrra/browser-logos/main/src/chrome/chrome_24x24.png "Chrome") | 116 | 116.0.5845.180 | Windows 10 | `chrome116` | | ![Chrome](https://raw.githubusercontent.com/alrra/browser-logos/main/src/chrome/chrome_24x24.png "Chrome") | 119 | 119.0.6045.199 | macOS Sonoma | `chrome119` | | ![Chrome](https://raw.githubusercontent.com/alrra/browser-logos/main/src/chrome/chrome_24x24.png "Chrome") | 120 | 120.0.6099.109 | macOS Sonoma | `chrome120` | | ![Chrome](https://raw.githubusercontent.com/alrra/browser-logos/main/src/chrome/chrome_24x24.png "Chrome") | 123 | 123.0.6312.124 | macOS Sonoma | `chrome123` | | ![Chrome](https://raw.githubusercontent.com/alrra/browser-logos/main/src/chrome/chrome_24x24.png "Chrome") | 124 | 124.0.6367.60 | macOS Sonoma | `chrome124` | | ![Chrome](https://raw.githubusercontent.com/alrra/browser-logos/main/src/chrome/chrome_24x24.png "Chrome") | 131 | 131.0.6778.86 | macOS Sonoma | `chrome131` | | ![Chrome](https://raw.githubusercontent.com/alrra/browser-logos/main/src/chrome/chrome_24x24.png "Chrome") | 131 | 131.0.6778.81 | Android 14 | `chrome131_android` | | ![Chrome](https://raw.githubusercontent.com/alrra/browser-logos/main/src/chrome/chrome_24x24.png "Chrome") | 133 | 133.0.6943.55 | macOS Sequoia | `chrome133a` | | ![Edge](https://raw.githubusercontent.com/alrra/browser-logos/main/src/edge/edge_24x24.png "Edge") | 99 | 99.0.1150.30 | Windows 10 | `edge99` | | ![Edge](https://raw.githubusercontent.com/alrra/browser-logos/main/src/edge/edge_24x24.png "Edge") | 101 | 101.0.1210.47 | Windows 10 | `edge101` | | ![Safari](https://github.com/alrra/browser-logos/blob/main/src/safari/safari_24x24.png "Safari") | 15.3 | 16612.4.9.1.8 | MacOS Big Sur | `safari15_3` | | ![Safari](https://github.com/alrra/browser-logos/blob/main/src/safari/safari_24x24.png "Safari") | 15.5 | 17613.2.7.1.8 | MacOS Monterey | `safari15_5` | | ![Safari](https://github.com/alrra/browser-logos/blob/main/src/safari/safari_24x24.png "Safari") | 17.0 | unclear | MacOS Sonoma | `safari17_0` | | ![Safari](https://github.com/alrra/browser-logos/blob/main/src/safari/safari_24x24.png "Safari") | 17.2 | unclear | iOS 17.2 | `safari17_2_ios` | | ![Safari](https://github.com/alrra/browser-logos/blob/main/src/safari/safari_24x24.png "Safari") | 18.0 | unclear | MacOS Sequoia | `safari18_0` | | ![Safari](https://github.com/alrra/browser-logos/blob/main/src/safari/safari_24x24.png "Safari") | 18.0 | unclear | iOS 18.0 | `safari18_0_ios` | | ![Firefox](https://github.com/alrra/browser-logos/blob/main/src/firefox/firefox_24x24.png "Firefox") | 133.0 | 133.0.3 | macOS Sonoma | `firefox133` | | ![Firefox](https://github.com/alrra/browser-logos/blob/main/src/firefox/firefox_24x24.png "Firefox") | 135.0 | 135.0.1 | macOS Sonoma | `firefox135` | ## Thanks This project is inspired by the following projects: + [curl_cffi](https://github.com/yifeikong/curl_cffi) - Python binding for curl-impersonate via cffi. A http client that can impersonate browser tls/ja3/http2 fingerprints. + [curl-impersonate](https://github.com/lwthiker/curl-impersonate) - A special build of curl that can impersonate Chrome & Firefox + [scrapy-playwright](https://github.com/scrapy-plugins/scrapy-playwright) - Playwright integration for Scrapy
text/markdown
Jalil SA (jxlil)
null
null
null
MIT
null
[ "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "License :: OSI Approved :: MIT License" ]
[]
https://github.com/jxlil/scrapy-impersonate
null
>=3.8
[]
[]
[]
[ "curl-cffi>=0.13.0", "scrapy>=2.12.0" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.14.2
2026-02-21T00:49:16.604694
scrapy_impersonate-1.6.3.tar.gz
6,735
04/3f/fa97b40cec0c601255bcdffc73a1916a70dd01bfc03620444861d779e529/scrapy_impersonate-1.6.3.tar.gz
source
sdist
null
false
5ec4bef72ec895719ca967799784080e
bf20edcb88b73b759d78bf0047b8acc91513bd44a8490cd600c5fa934fd0aa3c
043ffa97b40cec0c601255bcdffc73a1916a70dd01bfc03620444861d779e529
null
[ "LICENSE" ]
308
2.4
xu-agent-sdk
0.1.1
Python SDK for 1XU Trading Signals - Connect your autonomous agent to AI-powered signal pipeline
# xu-agent-sdk Python SDK for connecting autonomous trading agents to 1XU's AI-powered signal pipeline. ## Features - ⚡ **Real-time Signals** — AI-powered trading signals for Polymarket - 📡 **20+ Endpoints** — Full A2A protocol with REST API & webhooks - 📊 **Performance Tracking** — Report trades and track your P&L - 🔄 **Streaming** — Long-polling and webhook delivery - ✅ **On-chain Verification** — Prove your trades on-chain ## Installation ```bash pip install xu-agent-sdk ``` ## Quick Start ```python import asyncio from xu_agent_sdk import XuAgent async def main(): # Initialize with your API key agent = XuAgent(api_key="1xu_your_api_key_here") # Verify connection and get plan info info = await agent.verify_connection() print(f"Connected! Plan: {info['plan']}") # Get latest signals signals = await agent.get_signals(limit=5, min_confidence=0.6) for signal in signals: print(f"Signal: {signal.market}") print(f" Direction: {signal.direction}") print(f" Confidence: {signal.confidence:.1%}") print(f" Entry: {signal.entry_price:.3f}") print() await agent.close() asyncio.run(main()) ``` ## Getting an API Key 1. **Deposit USDC** at https://1xu.app and get your API key instantly 2. **$2/day** during 15-day trial, then **$10/day** 3. Cancel anytime ## Usage Examples ### Get Signals ```python # Get latest signals signals = await agent.get_signals(limit=10) # Filter by confidence signals = await agent.get_signals(min_confidence=0.7) # Filter by direction signals = await agent.get_signals(direction='yes') ``` ### Report Trades ```python # Report a trade execution result = await agent.report_trade( signal_id=signal.id, market_id=signal.market_id, direction='yes', size_usd=100.0, entry_price=0.65, tx_hash="0x..." # Optional, for verification ) trade_id = result['trade_id'] # Later, report the close await agent.report_close( trade_id=trade_id, exit_price=0.85, pnl_usd=30.77 # (0.85-0.65)/0.65 * 100 ) ``` ### Report Skips/Failures ```python # Skip a signal await agent.report_skip( signal_id=signal.id, reason="below_threshold" # or "no_liquidity", "already_positioned" ) # Report failure await agent.report_failure( signal_id=signal.id, reason="slippage_too_high" ) ``` ### Check Stats ```python # Your performance stats = await agent.get_my_stats() print(f"Win Rate: {stats['win_rate']:.1f}%") print(f"Total P&L: ${stats['total_pnl_usd']:.2f}") ``` ### Stream Signals (Long-polling) ```python async for signal in agent.stream_signals(min_confidence=0.6): print(f"New signal: {signal.market} -> {signal.direction}") # Execute your trading logic here ``` ### Register Webhook ```python # For real-time delivery await agent.register_webhook("https://your-agent.com/webhook") ``` ## Error Handling ```python from xu_agent_sdk import ( XuAuthError, XuRateLimitError, XuPaymentRequiredError, XuSignalLimitError ) try: signals = await agent.get_signals() except XuAuthError: print("Invalid API key") except XuRateLimitError as e: print(f"Rate limited, retry in {e.retry_after}s") except XuSignalLimitError as e: print(f"Daily limit reached, resets in {e.resets_in_seconds}s") except XuPaymentRequiredError as e: print(f"Deposit USDC required: {e.pricing}") ``` ## Webhook Payload When you register a webhook, signals are POSTed with this format: ```json { "event": "new_signal", "signal_id": "abc123", "timestamp": "2026-02-03T10:30:00Z", "data": { "market": "Will Bitcoin hit $100k by March 2026?", "market_id": "0x1234...", "direction": "yes", "confidence": 0.72, "suggested_size": "$50-100", "entry_price": 0.65 } } ``` Verify webhooks using the `X-1xu-Signature` header. ## Support - **Documentation**: https://1xu.app/docs/sdk - **Discord**: https://discord.gg/1xu - **Email**: dev@1xu.app ## License MIT
text/markdown
null
1xu Team <dev@1xu.app>
null
null
MIT
trading, polymarket, signals, crypto, autonomous-agent, a2a-protocol
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Topic :: Office/Business :: Financial :: Investment", "Framework :: AsyncIO" ]
[]
null
null
>=3.9
[]
[]
[]
[ "aiohttp>=3.8.0", "pytest>=7.0; extra == \"dev\"", "pytest-asyncio>=0.20; extra == \"dev\"", "black; extra == \"dev\"", "mypy; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://1xu.app", "Documentation, https://1xu.app/docs/sdk", "Repository, https://github.com/1xu-project/xu-agent-sdk", "Changelog, https://github.com/1xu-project/xu-agent-sdk/releases" ]
twine/6.2.0 CPython/3.9.6
2026-02-21T00:48:41.964321
xu_agent_sdk-0.1.1.tar.gz
10,919
a1/e0/5819fb097033dec273249f20f9884f7b2775eb47bde58e5db83009315f90/xu_agent_sdk-0.1.1.tar.gz
source
sdist
null
false
82262c2db4e752a55d91f0cbe4a4e259
a87e46ceba1c7f8ce93562613ea69feb4c26a37c5625f096d64e76608ba70934
a1e05819fb097033dec273249f20f9884f7b2775eb47bde58e5db83009315f90
null
[]
229
2.4
Geode-Hybrid
3.4.3
Hybrid remeshing Geode-solutions OpenGeode module
<h1 align="center">Geode-Hybrid_private<sup><i>by Geode-solutions</i></sup></h1> <h3 align="center">Template for creating your own OpenGeode private module</h3> <p align="center"> <img src="https://github.com/Geode-solutions/Geode-ModuleTemplate_private/workflows/CI/badge.svg" alt="Build Status"> <img src="https://github.com/Geode-solutions/Geode-ModuleTemplate_private/workflows/CD/badge.svg" alt="Deploy Status"> <img src="https://codecov.io/gh/Geode-solutions/Geode-ModuleTemplate_private/branch/master/graph/badge.svg" alt="Coverage Status"> <img src="https://img.shields.io/github/release/Geode-solutions/Geode-ModuleTemplate_private.svg" alt="Version"> </p> <p align="center"> <img src="https://img.shields.io/static/v1?label=Windows&logo=windows&logoColor=white&message=support&color=success" alt="Windows support"> <img src="https://img.shields.io/static/v1?label=Ubuntu&logo=Ubuntu&logoColor=white&message=support&color=success" alt="Ubuntu support"> <img src="https://img.shields.io/static/v1?label=Red%20Hat&logo=Red-Hat&logoColor=white&message=support&color=success" alt="Red Hat support"> </p> <p align="center"> <img src="https://img.shields.io/badge/C%2B%2B-11-blue.svg" alt="Language"> <img src="https://img.shields.io/badge/license-MIT-blue.svg" alt="License"> <img src="https://img.shields.io/badge/%20%20%F0%9F%93%A6%F0%9F%9A%80-semantic--release-e10079.svg" alt="Semantic-release"> <a href="https://geode-solutions.com/#slack"> <img src="https://opengeode-slack-invite.herokuapp.com/badge.svg" alt="Slack invite"> </a> Copyright (c) 2019 - 2026, Geode-solutions
text/markdown
null
Geode-solutions <contact@geode-solutions.com>
null
null
Proprietary
null
[]
[]
null
null
<3.13,>=3.9
[]
[]
[]
[ "geode-background==9.*,>=9.10.3", "geode-common==33.*,>=33.19.0", "geode-numerics==6.*,>=6.4.12", "geode-simplex==11.*,>=11.0.0", "opengeode-core==15.*,>=15.31.5", "opengeode-inspector==6.*,>=6.8.17" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.9.13
2026-02-21T00:48:07.411368
geode_hybrid-3.4.3-cp39-cp39-win_amd64.whl
1,792,180
f3/9b/ef87c97397ea1b096809ebf9a628bb000cd224c1b614cf9ebc8b1776a88d/geode_hybrid-3.4.3-cp39-cp39-win_amd64.whl
cp39
bdist_wheel
null
false
c060f417bf9f78f0b16986e6405c6e3c
2c903d7e87c5430cad670872e1115b3f01d6cc638c3afe7311e1faea1e7cc178
f39bef87c97397ea1b096809ebf9a628bb000cd224c1b614cf9ebc8b1776a88d
null
[]
0
2.4
zros
0.1.8
ZROS: ZeroMQ ROS-like framework
# ZROS: A fast, lightweight ROS-like library [![PyPI version](https://badge.fury.io/py/zros.svg)](https://badge.fury.io/py/zros) [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/licenses/MIT) [![Python versions](https://img.shields.io/pypi/pyversions/zros.svg)](https://pypi.org/project/zros) <div align="center"> <img src="https://github.com/user-attachments/assets/a9d1d0d0-23df-4f1c-9462-415914c68fa9" width="600" alt="zros_logo" /> </div> **ZROS** is a fast, lightweight [ROS](https://www.ros.org/)-like library designed to bring the ease of ROS-2 to Python projects that require minimal overhead and high performance, using [ZeroMQ](https://zeromq.org/) for fast, asynchronous communication. It provides a simple, pure Python alternative for robotic applications, computer vision pipelines, and distributed systems where a full ROS installation might be overkill. ## Key Features - **Fast & Lightweight:** Built on top of **ZeroMQ**, ensuring low-latency communication between nodes. - **ROS-like API:** Uses familiar concepts like `Node`, `Publisher`, `Subscriber`, `Timer`, and `spin()` making it easy for ROS-2 developers to adapt. - **No Complex Build System:** Pure Python. No `catkin_make`, no `colcon build`, no `source setup.bash`. Just run your Python scripts. - **Computer Vision Ready:** Includes a built-in `CvBridge` for seamless OpenCV image transport. ## Installation ### From PyPI (Recommended) ```bash uv venv uv pip install zros ``` ### From Source ```bash git clone https://github.com/juliodltv/zros.git cd zros uv sync ``` ## Quick Start ### Create a Publisher (publisher.py) ```python from zros import Node, CvBridge import cv2 class CameraPublisher(Node): def __init__(self): super().__init__("camera_pub") self.pub = self.create_publisher("video_topic") self.bridge = CvBridge() self.cap = cv2.VideoCapture(0) self.create_timer(1/60, self.timer_callback) def timer_callback(self): ret, frame = self.cap.read() if ret: # msg is a dictionary msg = { "image": self.bridge.cv2_to_msg(frame), "info": "My Camera Frame" } self.pub.publish(msg) if __name__ == "__main__": CameraPublisher().spin() ``` ### Create a Subscriber (subscriber.py) ```python from zros import Node, CvBridge import cv2 class VideoSubscriber(Node): def __init__(self): super().__init__("video_sub") self.bridge = CvBridge() self.create_subscriber("video_topic", self.callback) def callback(self, msg): img = self.bridge.msg_to_cv2(msg["image"]) # info = msg["info"] # print(info) cv2.imshow("Video", img) cv2.waitKey(1) if __name__ == "__main__": VideoSubscriber().spin() ``` ## Running Examples You can find another examples in the `examples/` directory. ```bash # Terminal 1 uv run zroscore # Terminal 2 uv run publisher.py # Terminal 3 uv run subscriber.py ``` ## Documentation For detailed usage instructions, please refer to the [ZROS Documentation](https://juliodltv.github.io/zros/). ## Citation ```bibtex @software{zros2026, author = {Julio De La Torre-Vanegas}, title = {ZROS: A fast, lightweight ZeroMQ ROS-like library}, year = {2026}, url = {https://github.com/juliodltv/zros} } ```
text/markdown
null
null
null
null
MIT
null
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "Intended Audience :: Science/Research", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.8", "Operating System :: OS Independent", "Topic :: Software Development :: Libraries :: Python Modules" ]
[]
null
null
>=3.8
[]
[]
[]
[ "opencv-python>=4.13.0.92", "pyzmq>=27.1.0", "mkdocs==1.6.1; extra == \"docs\"", "mkdocs-material==9.7.2; extra == \"docs\"", "mkdocstrings[python]==0.26.1; extra == \"docs\"" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:47:54.518930
zros-0.1.8.tar.gz
8,601
65/80/27d3eafcd03ba3720c78a55c27317d209f189b224dcb7ed7b1ff4ecb4738/zros-0.1.8.tar.gz
source
sdist
null
false
db39dcc1888be16b42709e925398fe9f
3952c44e5dec11b662ff3c190a7811c66a00d91f93249acde71ecf48c72e98d6
658027d3eafcd03ba3720c78a55c27317d209f189b224dcb7ed7b1ff4ecb4738
null
[ "LICENSE" ]
230
2.4
qrusty-pyclient
0.5.0
Python client wrapper for the qrusty API.
# qrusty_pyclient A Python client wrapper for the qrusty API. ## Features - Connect to a qrusty server - Publish, consume, ack, and purge messages - List and manage queues ## Installation ```bash pip install qrusty_pyclient ``` ## Usage ```python from qrusty_pyclient import QrustyClient client = QrustyClient(base_url="http://localhost:6784") client.create_queue(name="orders", ordering="MaxFirst", allow_duplicates=True) client.publish(queue="orders", priority=100, payload={"order_id": 123}) message = client.consume(queue="orders", consumer_id="worker-1") if message is not None: client.ack(queue="orders", message_id=message["id"], consumer_id="worker-1") ``` ## Development Don't forget your `~/.pypirc` file if you intend to publish to PyPI. ## License MIT
text/markdown
null
Gordon Greene <greeng3@obscure-reference.com>
null
null
null
null
[]
[]
null
null
>=3.7
[]
[]
[]
[ "requests" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.12.3
2026-02-21T00:47:26.005752
qrusty_pyclient-0.5.0.tar.gz
5,033
5e/af/7ce6f2057fe1a26b1cfafe1e8b956106d00e79e7f91922f386733bdd4b21/qrusty_pyclient-0.5.0.tar.gz
source
sdist
null
false
6b78394b981ca134a1a6517c01302af2
c659a4fb454ceb535893559966e1791c4b653270eb845a68c0bf23fda6e171b6
5eaf7ce6f2057fe1a26b1cfafe1e8b956106d00e79e7f91922f386733bdd4b21
MIT
[]
229
2.4
revenant
0.2.4
Cross-platform Python client for ARX CoSign electronic signatures via SOAP API
<p align="center"> <img src="icons/revenant-readme.png" width="128" alt="Revenant"> </p> # revenant (Python) [![CI](https://github.com/lobotomoe/revenant/actions/workflows/ci.yml/badge.svg)](https://github.com/lobotomoe/revenant/actions/workflows/ci.yml) [![pyright: strict](https://img.shields.io/badge/pyright-strict-blue)](https://github.com/microsoft/pyright) [![coverage: 96%](https://img.shields.io/badge/coverage-96%25-brightgreen)](pyproject.toml) [![code style: ruff](https://img.shields.io/badge/code%20style-ruff-000000.svg)](https://docs.astral.sh/ruff/) Cross-platform Python client for ARX CoSign electronic signatures via SOAP API. No Windows required. Server-specific settings (URL, TLS, identity discovery) are managed through **server profiles** — see [`src/revenant/config/profiles.py`](src/revenant/config/profiles.py). EKENG-specific details are documented in [`../docs/ekeng/`](../docs/ekeng/). For SOAP API technical details, see [`../docs/soap-api.md`](../docs/soap-api.md). ## Download Pre-built binaries are available on the [Releases](https://github.com/lobotomoe/revenant/releases) page: | Platform | CLI | GUI | |----------|-----|-----| | **macOS (Apple Silicon)** | `revenant-cli-macos-arm64` | `Revenant-gui-macos-arm64.dmg` | | **Linux (x64)** | `revenant-cli-linux-x64` | `Revenant-x86_64.AppImage` | | **Linux (ARM64)** | `revenant-cli-linux-arm64` | `Revenant-aarch64.AppImage` | | **Windows (x64)** | `revenant-windows-x64.zip` | `Revenant.msix` | ### Quick start (macOS) ```bash # Download and run CLI curl -LO https://github.com/lobotomoe/revenant/releases/latest/download/revenant-cli-macos-arm64 chmod +x revenant-cli-macos-arm64 ./revenant-cli-macos-arm64 setup ``` ### Quick start (Linux x64) ```bash # CLI curl -LO https://github.com/lobotomoe/revenant/releases/latest/download/revenant-cli-linux-x64 chmod +x revenant-cli-linux-x64 ./revenant-cli-linux-x64 setup # GUI (AppImage) curl -LO https://github.com/lobotomoe/revenant/releases/latest/download/Revenant-x86_64.AppImage chmod +x Revenant-x86_64.AppImage ./Revenant-x86_64.AppImage ``` ### Quick start (Linux ARM64) ```bash # CLI curl -LO https://github.com/lobotomoe/revenant/releases/latest/download/revenant-cli-linux-arm64 chmod +x revenant-cli-linux-arm64 ./revenant-cli-linux-arm64 setup # GUI (AppImage) curl -LO https://github.com/lobotomoe/revenant/releases/latest/download/Revenant-aarch64.AppImage chmod +x Revenant-aarch64.AppImage ./Revenant-aarch64.AppImage ``` ### Quick start (Windows) Download `revenant-windows-x64.zip` from [Releases](https://github.com/lobotomoe/revenant/releases), extract, and run: ```powershell Expand-Archive revenant-windows-x64.zip -DestinationPath revenant .\revenant\revenant.exe setup ``` Alternatively, install the GUI via `Revenant.msix` (double-click to install). ## Installation (pip) ```bash # Install directly from GitHub pip install "revenant @ git+https://github.com/lobotomoe/revenant.git#subdirectory=python" # Or clone and install in editable mode (for development) git clone https://github.com/lobotomoe/revenant cd revenant/python pip install -e "." # core functionality (includes pikepdf) pip install -e ".[secure]" # + secure credential storage (keyring) pip install -e ".[dev]" # + development tools (pytest, ruff, pyright) ``` ## Uninstall ```bash # PyPI pip uninstall revenant # Snap Store sudo snap remove revenant # Homebrew (macOS) brew uninstall lobotomoe/revenant/revenant ``` **Pre-built binaries:** delete the downloaded file (`revenant-cli-*`, `Revenant-*.AppImage`, etc.). **macOS DMG:** drag `Revenant.app` from `/Applications` to Trash. **Windows MSIX:** Settings > Apps > Revenant > Uninstall. **Remove configuration and saved credentials:** ```bash revenant reset # clears ~/.revenant/config.json ``` If you used keyring (secure credential storage), the password is stored in your system keychain. To remove it: - **macOS:** Keychain Access > search "revenant" > delete entry - **Linux:** `secret-tool clear service revenant` - **Windows:** Credential Manager > Windows Credentials > search "revenant" > remove ## Quick start (library) ```python import revenant # Sign with a built-in profile (handles URL, TLS, font automatically) signed_pdf = revenant.sign(pdf_bytes, "user", "pass", profile="ekeng") # Sign with a custom server URL signed_pdf = revenant.sign(pdf_bytes, "user", "pass", url="https://server/SAPIWS/DSS.asmx") # Use saved config (after `revenant setup`) signed_pdf = revenant.sign(pdf_bytes, "user", "pass") # Detached CMS/PKCS#7 signature cms_der = revenant.sign_detached(pdf_bytes, "user", "pass", profile="ekeng") ``` All functions raise typed exceptions (`AuthError`, `ServerError`, `TLSError`). See [Library usage](#library-usage) for the full low-level API. ## Usage ### Initial setup Configure your server, credentials, and signer identity: ```bash revenant setup # interactive wizard revenant setup --profile ekeng # skip server selection ``` The setup wizard walks you through: 1. **Choose server** — pick a built-in profile (e.g. EKENG) or enter a custom URL 2. **Ping server** — verify the endpoint is reachable (WSDL fetch, no auth) 3. **Enter credentials** — with lockout warnings for profiles that have them 4. **Discover identity** — automatically extracts your name/email/org from the server's signing certificate. Falls back to signed PDF extraction or manual entry. 5. **Save** — writes everything to `~/.revenant/config.json` You can re-run `revenant setup` at any time to reconfigure. ### GUI A graphical interface is available via tkinter (Python stdlib): ```bash revenant gui # if installed via pip revenant-gui # alternative entry point python -m revenant gui ``` The GUI provides file pickers for PDF/image/output, credential fields, position/page selectors, and a Sign button. Requires `tkinter` — if missing, the tool shows platform-specific install instructions (e.g. `brew install python-tk@3.13` on macOS). ### Sign a PDF (embedded — default) ```bash # Embedded signature — produces document_signed.pdf revenant sign document.pdf # Custom output path revenant sign document.pdf -o signed.pdf # Sign multiple files revenant sign *.pdf # Detached .p7s signature instead revenant sign document.pdf --detached # Preview what would be done (no actual signing) revenant sign document.pdf --dry-run # Specify page and position revenant sign document.pdf --page 1 --position top-left revenant sign document.pdf --page last --position bottom-center # Add signature image (PNG or JPEG, scaled to fit field, left side) revenant sign document.pdf --image signature.png # Armenian-script signature appearance revenant sign document.pdf --font ghea-grapalat ``` **Page numbering:** CLI uses 1-based pages (`--page 1` = first page). Use `first`, `last`, or a number. **Fonts:** `noto-sans` (default, Latin/Cyrillic), `ghea-grapalat` (Armenian), `ghea-mariam` (Armenian serif). The EKENG profile defaults to `ghea-grapalat`. **Signature image:** PNG or JPEG. The image is scaled proportionally to fit the left side of the signature field. Recommended: transparent PNG, around 200x100px. ### Verify a signature ```bash # Verify using CA root cert (auto-detected from certs/ directory) revenant verify document.pdf # Specify signature file explicitly revenant verify document.pdf -s document.pdf.p7s ``` ### Check an embedded signature ```bash revenant check signed.pdf ``` ### Inspect a detached signature ```bash revenant info document.pdf.p7s ``` ### Manage configuration ```bash revenant logout # clear credentials + identity, keep server config revenant reset # clear everything (server, credentials, identity) ``` `logout` preserves the server URL and profile so you can re-authenticate with `revenant setup` without reconfiguring the server. `reset` removes all configuration from `~/.revenant/config.json`. ### Output modes By default, the tool produces **embedded PDF signatures** (visible signature field in the PDF). Use `--detached` for detached CMS/PKCS#7 `.p7s` files. Embedded signing uses a **true incremental update** — the original PDF bytes are preserved exactly, with signature objects appended after `%%EOF`. pikepdf is used read-only (page dimensions, object graph) and never rewrites the PDF. Embedded signing includes **automatic post-sign verification** — after inserting the CMS, the tool re-reads the PDF and confirms the ByteRange hash matches what was sent to the server. If verification fails, the file is not saved. Embedded signatures include a **visual appearance stream** — fields configured per server profile (name, ID, date, etc.) are stacked vertically. With an optional signature image, the image appears on the left. Configure signer identity via `revenant setup`. The detached `.p7s` file can be verified with: - `openssl cms -verify -inform DER -in doc.pdf.p7s -content doc.pdf -CAfile certs/ca_root.pem` - Any PKCS#7/CMS-compatible verification tool ## Library usage The package can be used as a Python library — the CLI is just a thin wrapper. ```python import revenant from revenant.network.soap_transport import SoapSigningTransport transport = SoapSigningTransport("https://ca.gov.am:8080/SAPIWS/DSS.asmx") # Embedded signature (visible field in the PDF, requires pikepdf) # Library uses 0-based pages (page=0 = first page) signed_pdf = revenant.sign_pdf_embedded( pdf_bytes, transport, "user", "pass", timeout=120, page=0, x=350, y=50, w=200, h=70, name="Signer Name", reason="Approved", ) # Detached CMS/PKCS#7 signature cms_der = revenant.sign_pdf_detached(pdf_bytes, transport, "user", "pass") # Sign a raw SHA-1 hash (for custom workflows) import hashlib digest = hashlib.sha1(data).digest() cms_der = revenant.sign_hash(digest, transport, "user", "pass") # Sign arbitrary data (server computes the hash) cms_der = revenant.sign_data(raw_bytes, transport, "user", "pass") ``` ### Verification ```python # Verify the last embedded signature result = revenant.verify_embedded_signature(signed_pdf) print(result.valid, result.details) # Verify ALL signatures in a multi-signed PDF results = revenant.verify_all_embedded_signatures(signed_pdf) for r in results: print(r.signer, r.valid) ``` ### Signature positioning ```python # Available position presets print(revenant.POSITION_PRESETS) # {'bottom-right', 'top-right', 'bottom-left', 'top-left', 'bottom-center'} # Resolve aliases (e.g. "br" -> "bottom-right") pos = revenant.resolve_position("br") ``` ### Signature options `EmbeddedSignatureOptions` bundles all appearance and positioning parameters: ```python from revenant import EmbeddedSignatureOptions opts = EmbeddedSignatureOptions( page="last", # 0-based int, "first", or "last" position="bottom-right", # preset name (ignored when x/y are set) x=350, y=50, # manual coordinates (PDF points, origin=bottom-left) w=200, h=70, # field dimensions in PDF points reason="Approved", # signature reason string name="Signer Name", # signer display name image_path="sig.png", # optional PNG/JPEG signature image fields=["Name", "Date"], # custom appearance field strings visible=True, # False for invisible signatures font="noto-sans", # "noto-sans", "ghea-grapalat", or "ghea-mariam" ) signed = revenant.sign_pdf_embedded(pdf, transport, user, pw, options=opts) ``` ### Utilities ```python # Get configured signer name (from ~/.revenant/config.json) name = revenant.get_signer_name() # returns str | None ``` ### Error handling All functions raise typed exceptions from a hierarchy rooted at `RevenantError`: ``` RevenantError (base) ├── AuthError -- wrong credentials, account locked ├── ServerError -- server returned an error response ├── TLSError -- connection/TLS issues (.retryable flag) ├── PDFError -- invalid PDF structure, parse failures ├── ConfigError -- missing or malformed configuration └── CertificateError -- certificate parsing/extraction errors ``` ```python from revenant import AuthError, ServerError, TLSError, PDFError try: revenant.sign_pdf_embedded(pdf, transport, user, password) except AuthError: print("Wrong credentials or account locked") except TLSError as e: if e.retryable: print("Transient connection error, retry later") else: print(f"TLS configuration issue: {e}") except ServerError as e: print(f"Server error: {e}") except PDFError as e: print(f"Invalid PDF: {e}") ``` ### API stability This project follows semver. The public API (`revenant.__all__`) is stable from 1.0. Pre-1.0 releases may have breaking changes between minor versions. ## Server profiles Server-specific settings are managed through `ServerProfile` objects. The EKENG profile is built-in; custom servers are created at setup time. ### Built-in profile: EKENG - URL: `https://ca.gov.am:8080/SAPIWS/DSS.asmx` - TLS: Legacy TLSv1.0 / RC4-MD5 (auto-detected) - Account lockout: 5 failed attempts - Font: `ghea-grapalat` (Armenian) - Identity: extracted from signing certificate (name, SSN, email) ### Custom servers Run `revenant setup` and choose "Custom URL" to configure any CoSign server. The tool auto-detects whether the server requires legacy TLS on first connection. `ServerProfile` fields (defined in [`src/revenant/config/profiles.py`](src/revenant/config/profiles.py)): | Field | Type | Description | |-------|------|-------------| | `name` | `str` | Profile identifier | | `url` | `str` | SOAP endpoint URL (HTTPS only) | | `timeout` | `int` | Request timeout in seconds (default: 120) | | `legacy_tls` | `bool` | Force TLSv1.0/RC4 mode (default: auto-detect) | | `identity_methods` | `tuple` | Discovery methods: `"server"`, `"manual"` | | `ca_cert_markers` | `tuple` | Strings to identify CA certificates for filtering | | `max_auth_attempts` | `int` | Lockout threshold (0 = no lockout warning) | | `cert_fields` | `tuple` | Certificate fields for identity extraction | | `sig_fields` | `tuple` | Fields for signature visual appearance | | `font` | `str` | Default font for signatures | ## Prerequisites - Python 3.10+ - `pikepdf` — for embedded PDF signatures (brings in `qpdf`, `Pillow`, `lxml`) - `asn1crypto` — certificate parsing (PKCS#7, X.509) - `tlslite-ng` — legacy TLS for servers requiring TLS 1.0 / RC4 (e.g. EKENG) - `defusedxml` — safe XML parsing for SOAP responses - `openssl` for the `verify` command (optional, for detached signature verification) - CoSign credentials (username + password) - Network access to the CoSign server All Python dependencies are installed automatically via `pip install`. ### Platform notes | | macOS | Linux | Windows | |---|---|---|---| | **Signing** (`sign`) | works out of the box | works out of the box | works out of the box | | **Embedded PDF** | pikepdf | pikepdf | pikepdf | | **GUI** (`revenant gui`) | `brew install python-tk` | `apt install python3-tk` | included with Python | | **verify** | openssl included | openssl included | requires OpenSSL install | The core Python code is fully cross-platform. TLS is handled transparently: standard servers use system HTTPS (`urllib`), while legacy servers (e.g. EKENG with TLSv1.0/RC4) are handled via `tlslite-ng` (pure Python, no native dependencies). The transport layer auto-detects TLS mode per host on first connection. See [`../docs/ekeng/`](../docs/ekeng/) for EKENG-specific details. ### Credentials Credentials are resolved in this order (first match wins): 1. **Environment variables** `REVENANT_USER` / `REVENANT_PASS` 2. **System keychain** via `keyring` (if installed) 3. **Saved config** in `~/.revenant/config.json` (saved during `revenant setup` or after first successful sign) 4. **Interactive prompt** (if none of the above) After a successful signing from an interactive prompt, the tool offers to save credentials for future use. #### Secure storage (recommended) Install with keyring support for secure credential storage: ```bash pip install revenant[secure] # or pip install keyring ``` When `keyring` is installed, passwords are stored in your system's secure credential store: - **macOS**: Keychain - **Linux**: Secret Service (GNOME Keyring, KWallet) - **Windows**: Windows Credential Manager The username is still saved in `~/.revenant/config.json`, but the password is stored securely in the system keychain. #### Fallback (plaintext) If `keyring` is not installed, credentials are stored in `~/.revenant/config.json` (permissions `0600`). **You will see a warning** when saving credentials without keyring: ``` WARNING: Password is stored in PLAINTEXT (file is chmod 600) For secure storage, install: pip install keyring ``` To clear saved credentials, remove `username`/`password` from the config file or delete it. ### Environment variables | Variable | Description | | ------------------ | ----------------------------------------------------------------- | | `REVENANT_USER` | CoSign username (overrides saved config) | | `REVENANT_PASS` | CoSign password (overrides saved config) | | `REVENANT_URL` | SOAP endpoint (overrides profile URL from `revenant setup`) | | `REVENANT_TIMEOUT` | Request timeout in seconds (default: 120) | | `REVENANT_NAME` | Signer display name (overrides config from `revenant setup`) | ## Development ```bash cd python/ pip install -e ".[dev]" # install with dev tools (pytest, ruff, pyright) pytest # run unit tests (no server needed) ruff check src/ # lint ruff format src/ # format pyright src/ # type check # Integration tests (require live server + credentials) REVENANT_USER=... REVENANT_PASS=... pytest -m integration ``` ## Building from source Build standalone binaries (CLI + GUI) from the Python source. Each platform uses a different toolchain. ### macOS (.app + DMG) Uses [py2app](https://py2app.readthedocs.io/) for a sandbox-compatible `.app` bundle. ```bash cd python/ # Install build dependencies pip install -e ".[build-mac]" # Optional: install create-dmg for a fancy DMG layout (Applications link, background image) brew install create-dmg # Build .app bundle + DMG python scripts/build.py mac # Build .app only (no DMG -- useful if you want to sign before creating the DMG) python scripts/build.py mac --no-dmg # Create DMG from an existing .app python scripts/build.py dmg ``` **Requires:** Python 3.10+, tkinter (`brew install python-tk@3.13`), Xcode Command Line Tools. **Output:** `dist/Revenant.app`, `dist/Revenant.dmg` ### Linux (CLI + GUI + AppImage) Uses [PyInstaller](https://pyinstaller.org/) for standalone one-file binaries. ```bash cd python/ # Install build dependencies pip install -e ".[build]" # Build GUI + CLI binaries (runs in parallel) python scripts/build.py all # Build only CLI or GUI python scripts/build.py cli python scripts/build.py gui # Create AppImage from the GUI binary (requires Linux) python scripts/build_appimage.py ``` **Requires:** Python 3.10+, tkinter (`apt install python3-tk` for GUI). **Output:** `dist/revenant` (CLI), `dist/revenant-gui` (GUI), `dist/Revenant-{arch}.AppImage` (e.g. `Revenant-x86_64.AppImage`, `Revenant-aarch64.AppImage`) ### Windows (CLI + GUI + MSIX) Uses [Nuitka](https://nuitka.net/) for standalone executables (avoids Windows Defender false positives from PyInstaller's extract-to-temp pattern). ```bash cd python # Install build dependencies pip install -e ".[build-win]" # Build GUI + CLI (sequential -- Nuitka shares a download cache) python scripts/build.py all # Build only CLI or GUI python scripts/build.py cli python scripts/build.py gui # Create MSIX package (requires Windows SDK for makeappx.exe) python scripts/build_msix.py ``` **Requires:** Python 3.10+, tkinter (included with Python on Windows), Windows SDK (for MSIX only). **Output:** `dist/revenant-standalone/` (folder with `revenant.exe` + `revenant-gui.exe`), `dist/Revenant.msix` ## EKENG-specific notes EKENG-specific behavior is documented in [`../docs/ekeng/`](../docs/ekeng/) and configured as the `ekeng` profile in [`src/revenant/config/profiles.py`](src/revenant/config/profiles.py). Key points: - Server: `ca.gov.am:8080` (TLSv1.0 / RC4-MD5) - Account lockout after 5 failed attempts - Signatures are accepted by the EKENG validator (`ekeng.am`) and e-request (`e-request.am`) ## Troubleshooting **`AuthError: Authentication failed`** -- Wrong username or password. If using EKENG, the account locks after 5 failed attempts. Wait or contact your administrator. **`TLSError: ...`** -- Server unreachable or TLS version mismatch. Check network access to the server. For EKENG, the server requires TLSv1.0/RC4 which is handled automatically by `tlslite-ng`. **`ServerError: ...`** -- The server rejected the request. Common causes: expired certificate, server maintenance, or unsupported document format. **`PDFError: ...`** -- The PDF is malformed, encrypted, or not a valid PDF file. Try re-saving the PDF from a different application. **Signature appearance looks wrong** -- Run `revenant setup` to reconfigure your signer identity. The signature fields (name, ID, date) come from the server profile configuration. **Validator rejects the signed PDF** -- See [`../docs/ekeng/`](../docs/ekeng/) for EKENG validator requirements. Common issues: missing `/Info` dictionary in the incremental update, or modified PDF bytes after signing. ## Known limitations - **SHA-1 only** — the server rejects SHA-256. This is a server-side limitation. - **Non-standard CMS OIDs** — the server returns `sha1WithRSAEncryption` as digestAlgorithm (wrong per RFC 5652). See [`../docs/verification.md`](../docs/verification.md). - **No timestamp (TSA)** — the WSDL defines timestamp options but the server ignores them.
text/markdown
Aleksandr Kraiz
null
null
null
null
revenant, arx, cosign, dsa, electronic-signature, pdf, soap, pkcs7, cms, oasis-dss
[ "Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Developers", "Intended Audience :: Legal Industry", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Security :: Cryptography", "Typing :: Typed" ]
[]
null
null
>=3.10
[]
[]
[]
[ "pikepdf<11,>=8.0", "asn1crypto<2,>=1.5", "tlslite-ng<0.9,>=0.8.0", "defusedxml<1,>=0.7", "keyring<26,>=21.0; extra == \"secure\"", "pytest<10,>=7.0; extra == \"dev\"", "pytest-cov<8,>=4.0; extra == \"dev\"", "ruff<1.0,>=0.9; extra == \"dev\"", "pyright<2.0,>=1.1; extra == \"dev\"", "pillow<13,>=12.1.1; extra == \"dev\"", "keyring<26,>=21.0; extra == \"dev\"", "fonttools<5,>=4.0; extra == \"dev\"", "pyinstaller<7,>=6.0; extra == \"build\"", "nuitka<5,>=2.0; extra == \"build-win\"", "ordered-set<5,>=4.0; extra == \"build-win\"", "zstandard<1,>=0.20; extra == \"build-win\"", "py2app<1,>=0.28; extra == \"build-mac\"", "keyring<26,>=21.0; extra == \"build-mac\"" ]
[]
[]
[]
[ "Homepage, https://github.com/lobotomoe/revenant", "Repository, https://github.com/lobotomoe/revenant", "Issues, https://github.com/lobotomoe/revenant/issues", "Documentation, https://github.com/lobotomoe/revenant/tree/main/docs", "Changelog, https://github.com/lobotomoe/revenant/blob/main/CHANGELOG.md" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:46:33.644451
revenant-0.2.4.tar.gz
321,728
7b/ee/54f48faa6b9af76808371266ae9f1e4fca59fa0a91fdc76641a1ee7ad6f7/revenant-0.2.4.tar.gz
source
sdist
null
false
f96cc2f7d607de1078759e5ef0fdb7bb
4c26626fd8eaf43621590b54b65b3971671508183cc1587a0796ff268ce9463f
7bee54f48faa6b9af76808371266ae9f1e4fca59fa0a91fdc76641a1ee7ad6f7
Apache-2.0
[ "LICENSE" ]
222
2.4
mcp-server-grist
0.2.1
MCP server for Grist API integration
![Grist-Logo.png](https://www.getgrist.com/wp-content/uploads/2023/03/Grist-Logo.png) # Grist MCP Server [![PyPI version](https://img.shields.io/pypi/v/mcp-server-grist.svg)](https://pypi.org/project/mcp-server-grist/) [![Python versions](https://img.shields.io/pypi/pyversions/mcp-server-grist.svg)](https://pypi.org/project/mcp-server-grist/) [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/licenses/MIT) Un serveur MCP (Model Context Protocol) pour interagir avec l'API Grist. Ce serveur permet d'accéder et de manipuler les données Grist directement depuis des modèles de langage comme Claude. ## Structure du projet ``` mcp-server-grist/ ├── src/ │ └── mcp_server_grist/ # Package principal │ ├── __init__.py # Point d'entrée du package │ ├── __main__.py # Support pour exécution en module │ ├── version.py # Gestion de version │ ├── main.py # Point d'entrée principal │ ├── server.py # Configuration du serveur MCP │ ├── client.py # Client API Grist │ ├── tools/ # Outils MCP organisés par catégorie │ └── models.py # Modèles de données Pydantic ├── tests/ # Tests unitaires et d'intégration ├── docs/ # Documentation détaillée ├── requirements.txt # Dépendances Python ├── pyproject.toml # Configuration moderne du package ├── Dockerfile # Configuration Docker ├── docker-compose.yml # Configuration multi-services ├── .env.template # Template pour variables d'environnement └── README.md # Documentation principale ``` ## Prérequis - Python 3.8+ - Une clé API Grist valide - Les packages Python suivants : `fastmcp`, `httpx`, `pydantic`, `python-dotenv` ## Utilisation à la volée ### Via uvx (recommandé) En utilisant uvx, l'environnement et le téléchargement des paquets se fait à la volée au moment de l'éxécution ```bash uvx mcp-server-grist ``` ### Utilisation avec votre Assitant IA favori qui supporte le protocole MCP La configuration en json est : ```json { "mcpServers": { "grist-server": { "disabled": false, "timeout": 60, "type": "stdio", "command": "uvx", "args": [ "mcp-server-grist" ], "env": { "GRIST_API_KEY": "ta_cle_API_GRIST" } } } } ``` ## Installation ### Via pip ```bash pip install mcp-server-grist ``` Après l'installation, vous pouvez exécuter le serveur avec : ```bash mcp-server-grist ``` ### Utilisation avec Claude Desktop Pour utiliser ce serveur MCP avec Claude Desktop, ajoutez la configuration suivante à votre fichier `mcp_servers.json` : ```json { "mcpServers": { "grist-mcp": { "command": "node", "args": [ "chemin/vers/npm-wrapper/bin/start.js" ], "env": { "GRIST_API_KEY": "votre_clé_api_grist", "GRIST_API_URL": "https://docs.getgrist.com/api" } } } } ``` Remplacez `chemin/vers/npm-wrapper/bin/start.js` par le chemin absolu vers le script `start.js` du wrapper Node.js inclus dans ce package. ### Installation en mode développement Pour contribuer ou personnaliser le serveur : ```bash # Cloner le repository git clone https://github.com/modelcontextprotocol/mcp-server-grist.git cd mcp-server-grist # Installer en mode développement pip install -e . # Lancer les tests python -m pytest tests ``` ### Via Docker Pour un déploiement rapide avec Docker : ```bash # Construire l'image docker build -t mcp/grist-mcp-server . # Exécuter le container docker run -it --rm \ -e GRIST_API_KEY=votre_clé_api \ -e GRIST_API_HOST=https://docs.getgrist.com/api \ mcp/grist-mcp-server ``` ### Via Docker Compose Pour déployer plusieurs modes de transport en parallèle : ```bash # Configurer les variables d'environnement cp .env.example .env # Éditer le fichier .env avec votre clé API # Lancer les services docker-compose up ``` ## Configuration ### Variables d'environnement Créez un fichier `.env` basé sur `.env.template` avec les variables suivantes : ``` GRIST_API_KEY=votre_clé_api GRIST_API_HOST=https://docs.getgrist.com/api LOG_LEVEL=INFO # Options: DEBUG, INFO, WARNING, ERROR, CRITICAL ``` Vous trouverez votre clé API dans les paramètres de votre compte Grist. ### Configuration avec Claude Desktop Ajoutez ceci à votre `claude_desktop_config.json` : #### Version Python ```json { "mcpServers": { "grist-mcp": { "command": "python", "args": [ "-m", "grist_mcp_server" ] } } } ``` #### Version Docker ```json { "mcpServers": { "grist-mcp": { "command": "docker", "args": [ "run", "--rm", "-i", "-e", "GRIST_API_KEY=votre_clé_api", "-e", "GRIST_API_HOST=https://docs.getgrist.com/api", "mcp/grist-mcp-server" ] } } } ``` ## Options de démarrage Le serveur prend en charge plusieurs modes de transport conformes au standard MCP : ### En mode module (recommandé) ```bash # Mode stdio (par défaut pour Claude) python -m mcp_server_grist --transport stdio # Mode HTTP streamable (pour intégration web) python -m mcp_server_grist --transport streamable-http --host 127.0.0.1 --port 8000 --path /mcp # Active le mode debug avec logging détaillé python -m mcp_server_grist --debug ``` ### Options supplémentaires ``` Options: --transport {stdio,streamable-http,sse} Type de transport à utiliser --host HOST Hôte pour les transports HTTP (défaut: 127.0.0.1) --port PORT Port pour les transports HTTP (défaut: 8000) --path PATH Chemin pour streamable-http (défaut: /mcp) --mount-path MOUNT_PATH Chemin pour SSE (défaut: /sse) --debug Active le mode debug --help Affiche l'aide ``` ### Sécurité des transports Pour les transports HTTP, nous recommandons : - Utiliser `127.0.0.1` (localhost) plutôt que `0.0.0.0` pour limiter l'accès au réseau local - Activer la validation d'origine (`validate_origin`) pour éviter les attaques DNS rebinding - Pour une exposition à Internet, utiliser un proxy inverse avec HTTPS ## Fonctionnalités - Accès aux données Grist directement depuis les modèles de langage - Liste des organisations, espaces de travail, documents, tables et colonnes - Gestion des enregistrements (création, lecture, mise à jour, suppression) - Filtrage et tri des données avec des capacités de requêtage avancées - Support des requêtes SQL (SELECT uniquement) - Authentification sécurisée via clé API - Gestion des accès utilisateurs - Export et téléchargement (SQLite, Excel, CSV) - Gestion des pièces jointes - Gestion des webhooks - Validation intelligente des formules ## Outils disponibles ### Gestion des organisations et documents - `list_organizations` : Liste les organisations - `describe_organization` : Obtient des informations détaillées sur une organisation - `modify_organization` : Modifie une organisation - `delete_organization` : Supprime une organisation - `list_workspaces` : Liste les espaces de travail dans une organisation - `describe_workspace` : Obtient des informations détaillées sur un espace de travail - `create_workspace` : Crée un nouvel espace de travail - `modify_workspace` : Modifie un espace de travail - `delete_workspace` : Supprime un espace de travail - `list_documents` : Liste les documents dans un espace de travail - `describe_document` : Obtient des informations détaillées sur un document - `create_document` : Crée un nouveau document - `modify_document` : Modifie un document - `delete_document` : Supprime un document - `move_document` : Déplace un document vers un autre espace de travail - `force_reload_document` : Force le rechargement d'un document - `delete_document_history` : Supprime l'historique d'un document ### Gestion des tables et colonnes - `list_tables` : Liste les tables dans un document - `create_table` : Crée une nouvelle table - `modify_table` : Modifie une table - `list_columns` : Liste les colonnes dans une table - `create_column` : Crée une nouvelle colonne - `create_column_with_feedback` : Crée une colonne avec validation et retour détaillé - `modify_column` : Modifie une colonne - `delete_column` : Supprime une colonne - `create_column_with_formula_safe` : Crée une colonne de formule avec validation - `get_formula_helpers` : Obtient de l'aide pour construire des formules - `validate_formula` : Valide une formule et suggère des corrections - `get_table_schema` : Obtient le schéma d'une table ### Manipulation des données - `list_records` : Liste les enregistrements avec tri et limite - `add_grist_records` : Ajoute des enregistrements - `add_grist_records_safe` : Ajoute des enregistrements avec validation - `update_grist_records` : Met à jour des enregistrements - `delete_grist_records` : Supprime des enregistrements ### Filtrage et requêtes SQL - `filter_sql_query` : Requête SQL optimisée pour le filtrage simple * Interface simplifiée pour les filtres courants * Support du tri et de la limitation * Conditions WHERE basiques - `execute_sql_query` : Requête SQL complexe * Requêtes SQL personnalisées * Support des JOIN et sous-requêtes * Paramètres et timeout configurables ### Gestion des accès - `list_organization_access` : Liste les utilisateurs ayant accès à une organisation - `modify_organization_access` : Modifie l'accès d'un utilisateur à une organisation - `list_workspace_access` : Liste les utilisateurs ayant accès à un espace de travail - `modify_workspace_access` : Modifie l'accès d'un utilisateur à un espace de travail - `list_document_access` : Liste les utilisateurs ayant accès à un document - `modify_document_access` : Modifie l'accès d'un utilisateur à un document ### Export et téléchargement - `download_document_sqlite` : Télécharge un document au format SQLite - `download_document_excel` : Télécharge un document au format Excel - `download_table_csv` : Télécharge une table au format CSV ### Gestion des pièces jointes - `list_attachments` : Liste les pièces jointes dans un document - `get_attachment_info` : Obtient des informations sur une pièce jointe - `download_attachment` : Télécharge une pièce jointe - `upload_attachment` : Téléverse une pièce jointe ### Gestion des webhooks - `list_webhooks` : Liste les webhooks d'un document - `create_webhook` : Crée un webhook - `modify_webhook` : Modifie un webhook - `delete_webhook` : Supprime un webhook - `clear_webhook_queue` : Vide la file d'attente des webhooks ## Exemples d'utilisation ```python # Liste des organisations orgs = await list_organizations() # Liste des espaces de travail workspaces = await list_workspaces(org_id=1) # Liste des documents docs = await list_documents(workspace_id=1) # Liste des tables tables = await list_tables(doc_id="abc123") # Liste des colonnes columns = await list_columns(doc_id="abc123", table_id="Table1") # Liste des enregistrements avec tri et limite records = await list_records( doc_id="abc123", table_id="Table1", sort="name", limit=10 ) # Filtrage simple avec filter_sql_query filtered_records = await filter_sql_query( doc_id="abc123", table_id="Table1", columns=["name", "age", "status"], where_conditions={ "organisation": "OPSIA", "status": "actif" }, order_by="name", limit=10 ) # Requête SQL complexe avec execute_sql_query sql_result = await execute_sql_query( doc_id="abc123", sql_query=""" SELECT t1.name, t1.age, t2.department FROM Table1 t1 JOIN Table2 t2 ON t1.id = t2.employee_id WHERE t1.status = ? AND t1.age > ? ORDER BY t1.name LIMIT ? """, parameters=["actif", 25, 10], timeout_ms=2000 ) # Ajout d'enregistrements new_records = await add_grist_records( doc_id="abc123", table_id="Table1", records=[{"name": "John", "age": 30}] ) # Mise à jour d'enregistrements updated_records = await update_grist_records( doc_id="abc123", table_id="Table1", records=[{"id": 1, "name": "John", "age": 31}] ) # Création d'une colonne de formule avec validation formula_column = await create_column_with_formula_safe( doc_id="abc123", table_id="Table1", column_label="Total", formula="$Prix * $Quantité", column_type="Numeric" ) # Téléchargement d'un document au format Excel excel_doc = await download_document_excel( doc_id="abc123", header_format="label" ) # Gestion des accès await modify_document_access( doc_id="abc123", user_email="utilisateur@exemple.com", access_level="editors" ) ``` ## Cas d'utilisation détaillés ### Navigation et exploration - `list_organizations`, `list_workspaces`, `list_documents`, `list_tables`, `list_columns` * Utilisez pour explorer la structure Grist et découvrir les données disponibles * Nécessaires pour obtenir les IDs avant d'exécuter des opérations spécifiques * Idéal pour la phase initiale d'analyse des données ### Requêtes et filtrage - `list_records` : Pour obtenir tous les enregistrements d'une table - `filter_sql_query` : Pour les filtres simples sur une seule table - `execute_sql_query` : Pour les requêtes complexes avec jointures et sous-requêtes ### Manipulation des données - `add_grist_records` et `add_grist_records_safe` : Pour ajouter des données avec ou sans validation - `update_grist_records` : Pour modifier des enregistrements existants - `delete_grist_records` : Pour supprimer des enregistrements ### Travail avec les formules - `get_formula_helpers` : Pour obtenir la syntaxe correcte des références de colonnes - `validate_formula` : Pour vérifier et corriger automatiquement les formules - `create_column_with_formula_safe` : Pour créer des colonnes calculées sécurisées ### Export et téléchargement - `download_document_sqlite`, `download_document_excel`, `download_table_csv` : Pour exporter les données - `download_attachment` : Pour récupérer les fichiers attachés ### Gestion des accès - `list_*_access` et `modify_*_access` : Pour administrer les permissions utilisateurs ### Intégrations externes - `create_webhook`, `modify_webhook` : Pour connecter Grist à d'autres services ## Cas d'utilisation Le serveur MCP Grist est conçu pour : - Analyser et résumer les données Grist - Créer, mettre à jour et supprimer des enregistrements programmatiquement - Construire des rapports et des visualisations - Répondre aux questions sur les données stockées - Connecter Grist avec des modèles de langage pour des requêtes en langage naturel - Automatiser les flux de travail impliquant des données Grist - Intégrer Grist à d'autres systèmes via webhooks ## Contribution Les contributions sont les bienvenues ! Voici comment contribuer : 1. Forkez le projet 2. Créez une branche pour votre fonctionnalité 3. Committez vos changements 4. Poussez vers la branche 5. Ouvrez une Pull Request ## Licence Ce serveur MCP est sous licence MIT.
text/markdown
null
MCP Contributors <info@example.com>
null
null
null
null
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "Topic :: Software Development :: Libraries :: Python Modules", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Operating System :: OS Independent", "Intended Audience :: Science/Research", "Intended Audience :: Information Technology", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Communications", "Topic :: Internet" ]
[]
null
null
>=3.10
[]
[]
[]
[ "fastmcp>=2.8.0", "mcp>=1.9.4", "httpx>=0.24.0", "pydantic>=2.0.0", "python-dotenv>=1.0.0" ]
[]
[]
[]
[ "Homepage, https://github.com/nic01asFr/mcp-server-grist", "Documentation, https://github.com/nic01asFr/mcp-server-grist/blob/main/README.md", "Bug Reports, https://github.com/nic01asFr/mcp-server-grist/issues", "Source, https://github.com/nic01asFr/mcp-server-grist" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:46:23.562091
mcp_server_grist-0.2.1.tar.gz
62,337
30/a5/2ecebf5f084545f2a0ee01e99bc3af77d9941baeef3fb41e4da12b091f7b/mcp_server_grist-0.2.1.tar.gz
source
sdist
null
false
ae3e57b5f5d9d028ecf0705af1d43e83
b32d7b2c791e3dab974387e7a3be25d34354589eb4483a36e49b8a5aae751d04
30a52ecebf5f084545f2a0ee01e99bc3af77d9941baeef3fb41e4da12b091f7b
MIT
[ "LICENSE" ]
224
2.1
cloud-files
6.2.2
Fast access to cloud storage and local FS.
[![PyPI version](https://badge.fury.io/py/cloud-files.svg)](https://badge.fury.io/py/cloud-files) [![Test Suite](https://github.com/seung-lab/cloud-files/workflows/Test%20Suite/badge.svg)](https://github.com/seung-lab/cloud-files/actions?query=workflow%3A%22Test+Suite%22) CloudFiles: Fast access to cloud storage and local FS. ======== ```python from cloudfiles import CloudFiles, CloudFile, dl results = dl(["gs://bucket/file1", "gs://bucket2/file2", ... ]) # shorthand cf = CloudFiles('gs://bucket', progress=True) # s3://, https://, and file:// also supported results = cf.get(['file1', 'file2', 'file3', ..., 'fileN']) # threaded results = cf.get(paths, parallel=2) # threaded and two processes file1 = cf['file1'] part = cf['file1', 0:30] # first 30 bytes of file1 cf.put('filename', content) cf.put_json('filename', content) cf.puts([{ 'path': 'filename', 'content': content, }, ... ]) # automatically threaded cf.puts(content, parallel=2) # threaded + two processes cf.puts(content, storage_class="NEARLINE") # apply vendor-specific storage class cf.put_jsons(...) # same as puts cf['filename'] = content for fname in cf.list(prefix='abc123'): print(fname) list(cf) # same as list(cf.list()) cf.delete('filename') del cf['filename'] cf.delete([ 'filename_1', 'filename_2', ... ]) # threaded cf.delete(paths, parallel=2) # threaded + two processes boolean = cf.exists('filename') results = cf.exists([ 'filename_1', ... ]) # threaded cf.move("a", "gs://bucket/b") cf.moves("gs://bucket/", [ ("a", "b") ]) cf.touch("example") cf.touch([ "example", "example2" ]) ### NOTE: FOR SINGLE FILES ### Note CloudFiles plural vs CloudFile singular ### The examples below are for CloudFile cf = CloudFile("gs://bucket/file1") info = cf.head() binary = cf.get() obj = cf.get_json() cf.put(binary) cf.put_json() cf[:30] # get first 30 bytes of file num_bytes = cf.size() # get size in bytes (also in head) exists = cf.exists() # true or false cf.delete() # deletes the file cf.touch() # create the file if it doesn't exist cf.move("gs://example/destination/directory") # copy then delete source cf.transfer_from("gs://example/source/file.txt") # copies file efficiently cf.transfer_to("gs://example/dest/file.txt") # copies file efficiently path = cf.join([ path1, path2, path3 ]) # use the appropriate path separator ``` CloudFiles was developed to access files from object storage without ever touching disk. The goal was to reliably and rapidly access a petabyte of image data broken down into tens to hundreds of millions of files being accessed in parallel across thousands of cores. CloudFiles has been used to processes dozens of images, many of which were in the hundreds of terabyte range. It has reliably read and written tens of billions of files to date. ## Highlights 1. Fast file access with transparent threading and optionally multi-process access w/ local file locking. 2. Google Cloud Storage, Amazon S3, local filesystems, and arbitrary web servers making hybrid or multi-cloud easy. 3. Robust to flaky network connections. Uses exponential random window retries to avoid network collisions on a large cluster. Validates md5 for gcs and s3. 4. gzip, brotli, bz2, zstd, and xz compression. 5. Supports HTTP Range reads. 6. Supports green threads, which are important for achieving maximum performance on virtualized servers. 7. High efficiency transfers that avoid compression/decompression cycles. 8. High speed gzip decompression using libdeflate (compared with zlib). 9. Bundled CLI tool. 10. Accepts iterator and generator input. 11. Resumable bulk transfers. 12. Supports composite parallel upload for GCS and multi-part upload for AWS S3. 13. Supports s3 and GCS internal copies to avoid unnecessary data movement. ## Installation ```bash pip install cloud-files pip install cloud-files[test] # to enable testing with pytest pip install cloud-files[monitoring] # enable plotting network performance ``` If you run into trouble installing dependenies, make sure you're using at least Python3.6 and you have updated pip. On Linux, some dependencies require manylinux2010 or manylinux2014 binaries which earlier versions of pip do not search for. MacOS, Linux, and Windows are supported platforms. ### Credentials You may wish to install credentials under `~/.cloudvolume/secrets`. CloudFiles is descended from CloudVolume, and for now we'll leave the same configuration structure in place. You need credentials only for the services you'll use. The local filesystem doesn't need any. Google Storage ([setup instructions here](https://github.com/seung-lab/cloud-volume/wiki/Setting-up-Google-Cloud-Storage)) will attempt to use default account credentials if no service account is provided. If neither of those two conditions apply, you need a service account credential. `google-secret.json` is a service account credential for Google Storage, `aws-secret.json` is a service account for S3, etc. You can support multiple projects at once by prefixing the bucket you are planning to access to the credential filename. `google-secret.json` will be your defaut service account, but if you also want to also access bucket ABC, you can provide `ABC-google-secret.json` and you'll have simultaneous access to your ordinary buckets and ABC. The secondary credentials are accessed on the basis of the bucket name, not the project name. ```bash mkdir -p ~/.cloudvolume/secrets/ mv aws-secret.json ~/.cloudvolume/secrets/ # needed for Amazon mv google-secret.json ~/.cloudvolume/secrets/ # needed for Google mv matrix-secret.json ~/.cloudvolume/secrets/ # needed for Matrix ``` #### `aws-secret.json` and `matrix-secret.json` Create an [IAM user service account](https://docs.aws.amazon.com/IAM/latest/UserGuide/id_users.html) that can read, write, and delete objects from at least one bucket. ```json { "AWS_ACCESS_KEY_ID": "$MY_AWS_ACCESS_KEY_ID", "AWS_SECRET_ACCESS_KEY": "$MY_SECRET_ACCESS_TOKEN", "AWS_DEFAULT_REGION": "$MY_AWS_REGION" // defaults to us-east-1 } ``` #### `google-secret.json` You can create the `google-secret.json` file [here](https://console.cloud.google.com/iam-admin/serviceaccounts). You don't need to manually fill in JSON by hand, the below example is provided to show you what the end result should look like. You should be able to read, write, and delete objects from at least one bucket. ```json { "type": "service_account", "project_id": "$YOUR_GOOGLE_PROJECT_ID", "private_key_id": "...", "private_key": "...", "client_email": "...", "client_id": "...", "auth_uri": "https://accounts.google.com/o/oauth2/auth", "token_uri": "https://accounts.google.com/o/oauth2/token", "auth_provider_x509_cert_url": "https://www.googleapis.com/oauth2/v1/certs", "client_x509_cert_url": "" } ``` ## API Documentation Note that the "Cloud Costs" mentioned below are current as of June 2020 and are subject to change. As of this writing, S3 and Google use identical cost structures for these operations. `CloudFile` is a more intuitive version of `CloudFiles` designed for managing single files instead of groups of files. See examples above. There is an analogus method for each `CloudFiles` method (where it makes sense). ### Constructor ```python # import gevent.monkey # uncomment when using green threads # gevent.monkey.patch_all(thread=False) from cloudfiles import CloudFiles cf = CloudFiles( cloudpath, progress=False, green=None, secrets=None, num_threads=20, use_https=False, endpoint=None, request_payer=None, composite_upload_threshold = int(1e8) ) # cloudpath examples: cf = CloudFiles('gs://bucket/') # google cloud storage cf = CloudFiles('s3://bucket/') # Amazon S3 cf = CloudFiles('s3://https://s3emulator.com/coolguy/') # alternate s3 endpoint cf = CloudFiles('file:///home/coolguy/') # local filesystem cf = CloudFiles('mem:///home/coolguy/') # in memory cf = CloudFiles('https://website.com/coolguy/') # arbitrary web server ``` * cloudpath: The path to the bucket you are accessing. The path is formatted as `$PROTOCOL://BUCKET/PATH`. Files will then be accessed relative to the path. The protocols supported are `gs` (GCS), `s3` (AWS S3), `file` (local FS), `mem` (RAM), and `http`/`https`. * progress: Whether to display a progress bar when processing multiple items simultaneously. * green: Use green threads. For this to work properly, you must uncomment the top two lines. Green threads are used automatically upon monkey patching if green is None. * secrets: Provide secrets dynamically rather than fetching from the credentials directory `$HOME/.cloudvolume/secrets`. * num_threads: Number of simultaneous requests to make. Usually 20 per core is pretty close to optimal unless file sizes are extreme. * use_https: `gs://` and `s3://` require credentials to access their files. However, each has a read-only https endpoint that sometimes requires no credentials. If True, automatically convert `gs://` to `https://storage.googleapis.com/` and `s3://` to `https://s3.amazonaws.com/`. * endpoint: (s3 only) provide an alternate endpoint than the official Amazon servers. This is useful for accessing the various S3 emulators offered by on-premises deployments of object storage. * request_payer: Specify the account that should be charged for requests towards the bucket, rather than the bucket owner. * `gs://`: `request_payer` can be any Google Cloud project id. Please refer to the documentation for [more information](https://cloud.google.com/storage/docs/requester-pays). * `s3://`: `request_payer` must be `"requester"`. The AWS account associated with the AWS_ACCESS_KEY_ID will be charged. Please refer to the documentation for [more information](https://docs.aws.amazon.com/AmazonS3/latest/userguide/RequesterPaysBuckets.html) The advantage of using `mem://` versus a `dict` has both the advantage of using identical interfaces in your code and it will use compression automatically. ### get / get_json ```python # Let 'filename' be the file b'hello world' binary = cf.get('filename') binary = cf['filename'] >> b'hello world' compressed_binary = cf.get('filename', raw=True) binaries = cf.get(['filename1', 'filename2']) >> [ { 'path': 'filename1', 'content': b'...', 'byte_range': (None, None), 'error': None }, { 'path': 'filename2', 'content': b'...', 'byte_range': (None, None), 'error': None } ] # total provides info for progress bar when using generators. binaries = cf.get(generator(), total=N) binary = cf.get({ 'path': 'filename', 'start': 0, 'end': 5 }) # fetches 5 bytes binary = cf['filename', 0:5] # only fetches 5 bytes binary = cf['filename'][0:5] # same result, fetches 11 bytes >> b'hello' # represents byte range 0-4 inclusive of filename binaries = cf[:100] # download the first 100 files # Get the TransmissionMonitor object that records # the flight time of each file. binaries, tm = cf.get(..., return_recording=True) ``` `get` supports several different styles of input. The simplest takes a scalar filename and returns the contents of that file. However, you can also specify lists of filenames, a byte range request, and lists of byte range requests. You can provide a generator or iterator as input as well. Order is not guaranteed. When more than one file is provided at once, the download will be threaded using preemptive or cooperative (green) threads depending on the `green` setting. If `progress` is set to true, a progress bar will be displayed that counts down the number of files to download. `get_json` is the same as get but it will parse the returned binary as JSON data encoded as utf8 and returns a dictionary. Order is guaranteed. Cloud Cost: Usually about $0.40 per million requests. ### put / puts / put_json / put_jsons ```python cf.put('filename', b'content') cf['filename'] = b'content' cf.put_json('digits', [1,2,3,4,5]) cf.puts([{ 'path': 'filename', 'content': b'...', 'content_type': 'application/octet-stream', 'compress': 'gzip', 'compression_level': 6, # parameter for gzip or brotli compressor 'cache_control': 'no-cache', }]) cf.puts([ (path, content), (path, content) ], compression='gzip') cf.put_jsons(...) # Get the TransmissionMonitor object that records the # flight times of each object. _, tm = cf.puts(..., return_recording=True) # Definition of put, put_json is identical def put( self, path, content, content_type=None, compress=None, compression_level=None, cache_control=None, raw=False ) # Definition of puts, put_jsons is identical def puts( self, files, content_type=None, compress=None, compression_level=None, cache_control=None, total=None, raw=False ) ``` The PUT operation is the most complex operation because it's so configurable. Sometimes you want one file, sometimes many. Sometimes you want to configure each file individually, sometimes you want to standardize a bulk upload. Sometimes it's binary data, but oftentimes it's JSON. We provide a simpler interface for uploading a single file `put` and `put_json` (singular) versus the interface for uploading possibly many files `puts` and `put_jsons` (plural). In order to upload many files at once (which is much faster due to threading), you need to minimally provide the `path` and `content` for each file. This can be done either as a dict containing those fields or as a tuple `(path, content)`. If dicts are used, the fields (if present) specified in the dict take precedence over the parameters of the function. You can mix tuples with dicts. The input to puts can be a scalar (a single dict or tuple) or an iterable such as a list, iterator, or generator. Cloud Cost: Usually about $5 per million files. ### delete ```python cf.delete('filename') cf.delete([ 'file1', 'file2', ... ]) del cf['filename'] ``` This will issue a delete request for each file specified in a threaded fashion. Cloud Cost: Usually free. ### exists ```python cf.exists('filename') >> True # or False cf.exists([ 'file1', 'file2', ... ]) >> { 'file1': True, 'file2': False, ... } ``` Scalar input results in a simple boolean output while iterable input returns a dictionary of input paths mapped to whether they exist. In iterable mode, a progress bar may be displayed and threading is utilized to improve performance. Cloud Cost: Usually about $0.40 per million requests. ### size ```python cf.size('filename') >>> 1337 # size in bytes cf.size([ 'file1', 'nonexistent', 'empty_file', ... ]) >>> { 'file1': 392, 'nonexistent': None, 'empty_file': 0, ... } ``` The output is the size of each file as it is stored in bytes. If the file doesn't exist, `None` is returned. Scalar input results in a simple boolean output while iterable input returns a dictionary of input paths mapped to whether they exist. In iterable mode, a progress bar may be displayed and threading is utilized to improve performance. Cloud Cost: Usually about $0.40 per million requests. ### list ```python cf.list() # returns generator list(cf) # same as list(cf.list()) cf.list(prefix="abc") cf.list(prefix="abc", flat=True) ``` Recall that in object storage, directories do not really exist and file paths are really a key-value mapping. The `list` operator will list everything under the `cloudpath` given in the constructor. The `prefix` operator allows you to efficiently filter some of the results. If `flat` is specified, the results will be filtered to return only a single "level" of the "directory" even though directories are fake. The entire set of all subdirectories will still need to be fetched. Cloud Cost: Usually about $5 per million requests, but each request might list 1000 files. The list operation will continuously issue list requests lazily as needed. ### transfer_to / transfer_from ```python cff = CloudFiles('file:///source_location') cfg = CloudFiles('gs://dest_location') # Transfer all files from local filesys to google cloud storage cfg.transfer_from(cff, block_size=64) # in blocks of 64 files cff.transfer_to(cfg, block_size=64) cff.transfer_to(cfg, block_size=64, reencode='br') # change encoding to brotli cfg[:] = cff # default block size 64 ``` Transfer semantics provide a simple way to perform bulk file transfers. Use the `block_size` parameter to adjust the number of files handled in a given pass. This can be important for preventing memory blow-up and reducing latency between batches. gs to gs and s3 to s3 transfers will occur within the cloud without looping through the executing client provided no reencoding is specified. #### resumable transfer ```python from cloudfiles import ResumableTransfer # .db b/c this is a sqlite database # that will be automatically created rt = ResumableTransfer("NAME_OF_JOB.db") # init should only be called once rt.init("file://source_location", "gs://dest_location") # This part can be interrupted and resumed rt.execute("NAME_OF_JOB.db") # If multiple transfer clients, the lease_msec # parameter must be specified to prevent conflicts. rt.execute("NAME_OF_JOB.db", lease_msec=30000) rt.close() # deletes NAME_OF_JOB.db ``` This is esentially a more durable version of `cp`. The transfer works by first loading a sqlite database with filenames, a "done" flag, and a lease time. Then clients can attach to the database and execute the transfer in batches. When multiple clients are used, a lease time must be set so that the database does not return the same set of files to each client (and is robust). This transfer type can also be accessed via the CLI. ```bash cloudfiles xfer init SOURCE DEST --db NAME_OF_JOB.db cloudfiles xfer execute NAME_OF_JOB.db # deletes db when done ``` ### composite upload (Google Cloud Storage) If a file is larger than 100MB (default), CloudFiles will split the file into 100MB parts and upload them as individual part files using the STANDARD storage class to minimize deletion costs. Once uploaded, the part files will be recursively merged in a tree 32 files at a time. After each merge, the part files will be deleted. The final file will have the default storage class for the bucket. If an upload is interrupted, the part files will remain and must be cleaned up. You can provide an open for binary reading file handle instead of a bytes object so that large files can be uploaded without overwhelming RAM. You can also adjust the composite threshold using `CloudFiles(..., composite_upload_threshold=int(2e8))` to for example, raise the threshold to 200MB. ### mutli-part upload (S3) If a file is larger than 100MB (default), the S3 service will use multi-part upload. ou can provide an open for binary reading file handle instead of a bytes object so that large files can be uploaded without overwhelming RAM. You can also adjust the composite threshold using `CloudFiles(..., composite_upload_threshold=int(2e8))` to for example, raise the threshold to 200MB. Unfinished upload parts remain on S3 (and cost money) unless you use a bucket lifecycle rule to remove them automatically. https://docs.aws.amazon.com/AmazonS3/latest/userguide/mpuoverview.html ### transcode ```python from cloudfiles.compression import transcode files = cf.get(...) for file in transcode(files, 'gzip'): file['content'] # gzipped file content regardless of source transcode(files, encoding='gzip', # any cf compatible compression scheme in_place=False, # modify the files in-place to save memory progress=True # progress bar ) ``` Sometimes we want to change the encoding type of a set of arbitrary files (often when moving them around to another storage system). `transcode` will take the output of `get` and transcode the resultant files into a new format. `transcode` respects the `raw` attribute which indicates that the contents are already compressed and will decompress them first before recompressing. If the input data are already compressed to the correct output encoding, it will simply pass it through without going through a decompression/recompression cycle. `transcode` returns a generator so that the transcoding can be done in a streaming manner. ## Network Robustness CloudFiles protects itself from network issues in several ways. First, it uses a connection pool to avoid needing to reestablish connections or exhausting the number of available sockets. Second, it uses an exponential random window backoff to retry failed connections and requests. The exponential backoff allows increasing breathing room for an overloaded server and the random window decorrelates independent attempts by a cluster. If the backoff was not growing, the retry attempts by a large cluster would be too rapid fire or inefficiently slow. If the attempts were not decorrellated, then regardless of the backoff, the servers will often all try again around the same time. We backoff seven times starting from 0.5 seconds to 60 seconds, doubling the random window each time. Third, for Google Cloud Storage (GCS) and S3 endpoints, we compute the md5 digest both sending and receiving to ensure data corruption did not occur in transit and that the server sent the full response. We cannot validate the digest for partial ("Range") reads. For [composite objects](https://cloud.google.com/storage/docs/composite-objects) (GCS) we can check the [crc32c](https://pypi.org/project/crc32c/) check-sum which catches transmission errors but not tampering (though MD5 isn't secure at all anymore). We are unable to perform validation for [multi-part uploads](https://docs.aws.amazon.com/AmazonS3/latest/dev/mpuoverview.html) (S3). Using custom encryption keys may also create validation problems. ## Local File Locking When accessing local files (`file://`), CloudFiles can use the `fasteners` library to perform locking so multi-process IO can be performed safely. These options have no effect on remote file access such as `gs://` or `s3://`. Lock files are by default stored in `$CLOUD_FILES_DIR/locks` (usually `~/.cloudfiles/locks`). Locking is enabled if: 1. `CloudFiles(..., locking=True)` 2. `CloudFiles(..., locking=None)` and a locking directory is set either via `CloudFiles(..., lock_dir=...)` or via the environment variable `CLOUD_FILES_LOCK_DIR`. Locking is always disabled if `locking=False`. This can be advantageous for performance reasons but may require careful design of access patterns to avoid reading a file that is being written to. You can check the current assigned locks directory: ```python cf.lock_dir ``` You can clear the lock dir of all locks with: ```python cf.clear_locks() ``` ## CloudFiles CLI Tool ```bash # list cloud and local directories cloudfiles ls gs://bucket-folder/ # parallel file transfer, no decompression cloudfiles -p 2 cp --progress -r s3://bkt/ gs://bkt2/ # change compression type to brotli cloudfiles cp -c br s3://bkt/file.txt gs://bkt2/ # decompress cloudfiles cp -c none s3://bkt/file.txt gs://bkt2/ # save chart of file flight times cloudfiles cp --flight-time s3://bkt/file.txt gs://bkt2/ # save a chart of estimated bandwidth usage from these files alone cloudfiles cp --io-rate s3://bkt/file.txt gs://bkt2/ # save a chart of measured bandwidth usage for the machine cloudfiles cp --machine-io-rate s3://bkt/file.txt gs://bkt2/ # move or rename files cloudfiles mv s3://bkt/file.txt gs://bkt2/ # create an empty file if not existing cloudfiles touch s3://bkt/empty.txt # pass from stdin (use "-" for source argument) find some_dir | cloudfiles cp - s3://bkt/ # resumable transfers cloudfiles xfer init SRC DEST --db JOBNAME.db cloudfiles xfer execute JOBNAME.db --progress # can quit and resume # Get human readable file sizes from anywhere cloudfiles du -shc ./tmp gs://bkt/dir s3://bkt/dir # remove files cloudfiles rm ./tmp gs://bkt/dir/file s3://bkt/dir/file # cat across services, -r for range reads cloudfiles cat ./tmp gs://bkt/dir/file s3://bkt/dir/file # verify a transfer was successful by comparing bytes and hashes cloudfiles verify ./my-data gs://bucket/my-data/ ``` ### `cp` Pros and Cons For the cp command, the bundled CLI tool has a number of advantages vs. `gsutil` when it comes to transfers. 1. No decompression of file transfers (unless you want to). 2. Can shift compression format. 3. Easily control the number of parallel processes. 4. Green threads make core utilization more efficient. 5. Optionally uses libdeflate for faster gzip decompression. It also has some disadvantages: 1. Doesn't support all commands. 2. File suffixes may be added to signify compression type on the local filesystem (e.g. `.gz`, `.br`, or `.zstd`). `cloudfiles ls` will list them without the extension and they will be converted into `Content-Encoding` on cloud storage. ### `ls` Generative Expressions For the `ls` command, we support (via the `-e` flag) simple generative expressions that enable querying multiple prefixes at once. A generative expression is denoted `[chars]` where `c`,`h`,`a`,`r`, & `s` will be inserted individually into the position where the expression appears. Multiple expressions are allowed and produce a cartesian product of resulting strings. This functionality is very limited at the moment but we intend to improve it. ```bash cloudfiles ls -e "gs://bucket/prefix[ab]" # equivalent to: # cloudfiles ls gs://bucket/prefixa # cloudfiles ls gs://bucket/prefixb ``` ### `alias` for Alternative S3 Endpoints You can set your own protocols for S3 compatible endpoints by creating dynamic or persistent aliases. CloudFiles comes with two official s3 endpoints that are important for the Seung Lab, `matrix://` and `tigerdata://` which point to Princeton S3 endpoints. Official aliases can't be overridden. To create a dynamic alias, you can use `cloudfiles.paths.add_alias` which will only affect the current process. To create a persistent alias that resides in `~/.cloudfiles/aliases.json`, you can use the CLI. ```bash cloudfiles alias add example s3://https://example.com/ # example:// cloudfiles alias ls # list all aliases cloudfiles alias rm example # remove example:// ``` The alias file is only accessed (and cached) if CloudFiles encounters an unknown protocol. If you stick to default protocols and use the syntax `s3://https://example.com/` for alternative endpoints, you can still use CloudFiles in environments without filesystem access. ## Performance Monitoring CloudFiles now comes with two tools inside of `cloudfiles.monitoring` for measuring the performance of transfer operations both via the CLI and the programatic interface. A `TransmissionMonitor` object is created during each download or upload (e.g. `cf.get` or `cf.puts`) call. You can access this object by using the `return_recording=True` flag. This object saves the flight times of each object along with its size in an interval tree datastructure. It comes with methods for estimating the peak bits per a second and can plot both time of flight and the estimated transfer rates (assuming the transfer is evenly divided over the flight of an object, an assumption that is not always true). This allows you to estimate the contribution of a given CloudFiles operation to a machine's network IO. ```python from cloudfiles import CloudFiles ... results, tm = cf.get([ ... some files ... ], return_recording=True) value = tm.peak_Mbps() # estimated peak transfer rate value = tm.total_Mbps() # estimated average transfer rate tm.plot_gantt() # time of flight chart tm.plot_histogram() # transfer rate chart ``` A second object, `IOSampler`, can sample the OS network counters using a background thread and provides a global view of the machine's network performance during the life of the transfer. It is enabled on the CLI for the `cp` command when the `--machine-io-rate` flag is enabled, but must be manually started programatically. This is to avoid accidentally starting unnecessary sampling threads. The samples are accumulated into a circular buffer, so make sure to set the buffer length long enough for your points of interest to be captured. ```python from cloudfiles.monitoring import IOSampler sampler = IOSampler(buffer_sec=600, interval=0.25) sampler.start_sampling() ... sampler.stop_sampling() sampler.plot_histogram() ``` ## Credits CloudFiles is derived from the [CloudVolume.Storage](https://github.com/seung-lab/cloud-volume/tree/master/cloudvolume/storage) system. Storage was initially created by William Silversmith and Ignacio Tartavull. It was maintained and improved by William Silversmith and includes improvements by Nico Kemnitz (extremely fast exists), Ben Falk (brotli), and Ran Lu (local file locking). Manuel Castro added the ability to chose cloud storage class. Thanks to the anonymous author from https://teppen.io/ for their s3 etag validation code.
text/markdown
William Silversmith
ws9@princeton.edu
null
null
BSD-3-Clause
null
[ "Intended Audience :: Developers", "Development Status :: 4 - Beta", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Software Development :: Libraries :: Python Modules" ]
[]
https://github.com/seung-lab/cloud-files/
null
<4.0,>=3.9
[]
[]
[]
[ "boto3>=1.4.7", "brotli", "crc32c", "chardet>=3.0.4", "click", "deflate>=0.2.0", "gevent", "google-auth>=1.10.0", "google-cloud-core>=1.1.0", "google-cloud-storage>=1.31.1", "google-crc32c>=1.0.0", "intervaltree", "numpy", "orjson", "pathos", "protobuf>=3.3.0", "requests>=2.22.0", "six>=1.14.0", "tenacity>=4.10.0", "tqdm", "urllib3>=1.26.3", "zstandard", "rsa>=4.7.2", "fasteners", "pytest; extra == \"test\"", "moto>=5; extra == \"test\"", "numpy; extra == \"numpy\"", "psutil; extra == \"monitoring\"", "intervaltree; extra == \"monitoring\"", "matplotlib; extra == \"monitoring\"", "lxml; extra == \"apache\"" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.11.3
2026-02-21T00:44:10.219836
cloud_files-6.2.2.tar.gz
99,328
f1/c1/559547810240ffbdc014587bed5eb03c121c2b05fae9c49ae9383b2a9707/cloud_files-6.2.2.tar.gz
source
sdist
null
false
0af065fbd40acb88ced7b0d0c5ea5b83
27ac5b1325fdc0f64124bc518c49d66e2c279ea948daa32e416c9b9abf1a435b
f1c1559547810240ffbdc014587bed5eb03c121c2b05fae9c49ae9383b2a9707
null
[]
485
2.4
pipecat-ai
0.0.103
An open source framework for voice (and multimodal) assistants
<h1><div align="center"> <img alt="pipecat" width="300px" height="auto" src="https://raw.githubusercontent.com/pipecat-ai/pipecat/main/pipecat.png"> </div></h1> [![PyPI](https://img.shields.io/pypi/v/pipecat-ai)](https://pypi.org/project/pipecat-ai) ![Tests](https://github.com/pipecat-ai/pipecat/actions/workflows/tests.yaml/badge.svg) [![codecov](https://codecov.io/gh/pipecat-ai/pipecat/graph/badge.svg?token=LNVUIVO4Y9)](https://codecov.io/gh/pipecat-ai/pipecat) [![Docs](https://img.shields.io/badge/Documentation-blue)](https://docs.pipecat.ai) [![Discord](https://img.shields.io/discord/1239284677165056021)](https://discord.gg/pipecat) [![Ask DeepWiki](https://deepwiki.com/badge.svg)](https://deepwiki.com/pipecat-ai/pipecat) # 🎙️ Pipecat: Real-Time Voice & Multimodal AI Agents **Pipecat** is an open-source Python framework for building real-time voice and multimodal conversational agents. Orchestrate audio and video, AI services, different transports, and conversation pipelines effortlessly—so you can focus on what makes your agent unique. > Want to dive right in? Try the [quickstart](https://docs.pipecat.ai/getting-started/quickstart). ## 🚀 What You Can Build - **Voice Assistants** – natural, streaming conversations with AI - **AI Companions** – coaches, meeting assistants, characters - **Multimodal Interfaces** – voice, video, images, and more - **Interactive Storytelling** – creative tools with generative media - **Business Agents** – customer intake, support bots, guided flows - **Complex Dialog Systems** – design logic with structured conversations ## 🧠 Why Pipecat? - **Voice-first**: Integrates speech recognition, text-to-speech, and conversation handling - **Pluggable**: Supports many AI services and tools - **Composable Pipelines**: Build complex behavior from modular components - **Real-Time**: Ultra-low latency interaction with different transports (e.g. WebSockets or WebRTC) ## 🌐 Pipecat Ecosystem ### 📱 Client SDKs Building client applications? You can connect to Pipecat from any platform using our official SDKs: <a href="https://docs.pipecat.ai/client/js/introduction">JavaScript</a> | <a href="https://docs.pipecat.ai/client/react/introduction">React</a> | <a href="https://docs.pipecat.ai/client/react-native/introduction">React Native</a> | <a href="https://docs.pipecat.ai/client/ios/introduction">Swift</a> | <a href="https://docs.pipecat.ai/client/android/introduction">Kotlin</a> | <a href="https://docs.pipecat.ai/client/c++/introduction">C++</a> | <a href="https://github.com/pipecat-ai/pipecat-esp32">ESP32</a> ### 🧭 Structured conversations Looking to build structured conversations? Check out [Pipecat Flows](https://github.com/pipecat-ai/pipecat-flows) for managing complex conversational states and transitions. ### 🪄 Beautiful UIs Want to build beautiful and engaging experiences? Checkout the [Voice UI Kit](https://github.com/pipecat-ai/voice-ui-kit), a collection of components, hooks and templates for building voice AI applications quickly. ### 🛠️ Create and deploy projects Create a new project in under a minute with the [Pipecat CLI](https://github.com/pipecat-ai/pipecat-cli). Then use the CLI to monitor and deploy your agent to production. ### 🔍 Debugging Looking for help debugging your pipeline and processors? Check out [Whisker](https://github.com/pipecat-ai/whisker), a real-time Pipecat debugger. ### 🖥️ Terminal Love terminal applications? Check out [Tail](https://github.com/pipecat-ai/tail), a terminal dashboard for Pipecat. ### 📺️ Pipecat TV Channel Catch new features, interviews, and how-tos on our [Pipecat TV](https://www.youtube.com/playlist?list=PLzU2zoMTQIHjqC3v4q2XVSR3hGSzwKFwH) channel. ## 🎬 See it in action <p float="left"> <a href="https://github.com/pipecat-ai/pipecat-examples/tree/main/simple-chatbot"><img src="https://raw.githubusercontent.com/pipecat-ai/pipecat-examples/main/simple-chatbot/image.png" width="400" /></a>&nbsp; <a href="https://github.com/pipecat-ai/pipecat-examples/tree/main/storytelling-chatbot"><img src="https://raw.githubusercontent.com/pipecat-ai/pipecat-examples/main/storytelling-chatbot/image.png" width="400" /></a> <br/> <a href="https://github.com/pipecat-ai/pipecat-examples/tree/main/translation-chatbot"><img src="https://raw.githubusercontent.com/pipecat-ai/pipecat-examples/main/translation-chatbot/image.png" width="400" /></a>&nbsp; <a href="https://github.com/pipecat-ai/pipecat/blob/main/examples/foundational/12-describe-video.py"><img src="https://github.com/pipecat-ai/pipecat/blob/main/examples/foundational/assets/moondream.png" width="400" /></a> </p> ## 🧩 Available services | Category | Services | | ------------------- | ----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | | Speech-to-Text | [AssemblyAI](https://docs.pipecat.ai/server/services/stt/assemblyai), [AWS](https://docs.pipecat.ai/server/services/stt/aws), [Azure](https://docs.pipecat.ai/server/services/stt/azure), [Cartesia](https://docs.pipecat.ai/server/services/stt/cartesia), [Deepgram](https://docs.pipecat.ai/server/services/stt/deepgram), [ElevenLabs](https://docs.pipecat.ai/server/services/stt/elevenlabs), [Fal Wizper](https://docs.pipecat.ai/server/services/stt/fal), [Gladia](https://docs.pipecat.ai/server/services/stt/gladia), [Google](https://docs.pipecat.ai/server/services/stt/google), [Gradium](https://docs.pipecat.ai/server/services/stt/gradium), [Groq (Whisper)](https://docs.pipecat.ai/server/services/stt/groq), [Hathora](https://docs.pipecat.ai/server/services/stt/hathora), [NVIDIA Riva](https://docs.pipecat.ai/server/services/stt/riva), [OpenAI (Whisper)](https://docs.pipecat.ai/server/services/stt/openai), [SambaNova (Whisper)](https://docs.pipecat.ai/server/services/stt/sambanova), [Sarvam](https://docs.pipecat.ai/server/services/stt/sarvam), [Soniox](https://docs.pipecat.ai/server/services/stt/soniox), [Speechmatics](https://docs.pipecat.ai/server/services/stt/speechmatics), [Whisper](https://docs.pipecat.ai/server/services/stt/whisper) | | LLMs | [Anthropic](https://docs.pipecat.ai/server/services/llm/anthropic), [AWS](https://docs.pipecat.ai/server/services/llm/aws), [Azure](https://docs.pipecat.ai/server/services/llm/azure), [Cerebras](https://docs.pipecat.ai/server/services/llm/cerebras), [DeepSeek](https://docs.pipecat.ai/server/services/llm/deepseek), [Fireworks AI](https://docs.pipecat.ai/server/services/llm/fireworks), [Gemini](https://docs.pipecat.ai/server/services/llm/gemini), [Grok](https://docs.pipecat.ai/server/services/llm/grok), [Groq](https://docs.pipecat.ai/server/services/llm/groq), [Mistral](https://docs.pipecat.ai/server/services/llm/mistral), [NVIDIA NIM](https://docs.pipecat.ai/server/services/llm/nim), [Ollama](https://docs.pipecat.ai/server/services/llm/ollama), [OpenAI](https://docs.pipecat.ai/server/services/llm/openai), [OpenRouter](https://docs.pipecat.ai/server/services/llm/openrouter), [Perplexity](https://docs.pipecat.ai/server/services/llm/perplexity), [Qwen](https://docs.pipecat.ai/server/services/llm/qwen), [SambaNova](https://docs.pipecat.ai/server/services/llm/sambanova) [Together AI](https://docs.pipecat.ai/server/services/llm/together) | | Text-to-Speech | [Async](https://docs.pipecat.ai/server/services/tts/asyncai), [AWS](https://docs.pipecat.ai/server/services/tts/aws), [Azure](https://docs.pipecat.ai/server/services/tts/azure), [Camb AI](https://docs.pipecat.ai/server/services/tts/camb), [Cartesia](https://docs.pipecat.ai/server/services/tts/cartesia), [Deepgram](https://docs.pipecat.ai/server/services/tts/deepgram), [ElevenLabs](https://docs.pipecat.ai/server/services/tts/elevenlabs), [Fish](https://docs.pipecat.ai/server/services/tts/fish), [Google](https://docs.pipecat.ai/server/services/tts/google), [Gradium](https://docs.pipecat.ai/server/services/tts/gradium), [Groq](https://docs.pipecat.ai/server/services/tts/groq), [Hathora](https://docs.pipecat.ai/server/services/tts/hathora), [Hume](https://docs.pipecat.ai/server/services/tts/hume), [Inworld](https://docs.pipecat.ai/server/services/tts/inworld), [LMNT](https://docs.pipecat.ai/server/services/tts/lmnt), [MiniMax](https://docs.pipecat.ai/server/services/tts/minimax), [Neuphonic](https://docs.pipecat.ai/server/services/tts/neuphonic), [NVIDIA Riva](https://docs.pipecat.ai/server/services/tts/riva), [OpenAI](https://docs.pipecat.ai/server/services/tts/openai), [Piper](https://docs.pipecat.ai/server/services/tts/piper), [PlayHT](https://docs.pipecat.ai/server/services/tts/playht), [Resemble](https://docs.pipecat.ai/server/services/tts/resemble), [Rime](https://docs.pipecat.ai/server/services/tts/rime), [Sarvam](https://docs.pipecat.ai/server/services/tts/sarvam), [Speechmatics](https://docs.pipecat.ai/server/services/tts/speechmatics), [XTTS](https://docs.pipecat.ai/server/services/tts/xtts) | | Speech-to-Speech | [AWS Nova Sonic](https://docs.pipecat.ai/server/services/s2s/aws), [Gemini Multimodal Live](https://docs.pipecat.ai/server/services/s2s/gemini), [Grok Voice Agent](https://docs.pipecat.ai/server/services/s2s/grok), [OpenAI Realtime](https://docs.pipecat.ai/server/services/s2s/openai), [Ultravox](https://docs.pipecat.ai/server/services/s2s/ultravox), | | Transport | [Daily (WebRTC)](https://docs.pipecat.ai/server/services/transport/daily), [FastAPI Websocket](https://docs.pipecat.ai/server/services/transport/fastapi-websocket), [SmallWebRTCTransport](https://docs.pipecat.ai/server/services/transport/small-webrtc), [WebSocket Server](https://docs.pipecat.ai/server/services/transport/websocket-server), Local | | Serializers | [Exotel](https://docs.pipecat.ai/server/utilities/serializers/exotel), [Plivo](https://docs.pipecat.ai/server/utilities/serializers/plivo), [Twilio](https://docs.pipecat.ai/server/utilities/serializers/twilio), [Telnyx](https://docs.pipecat.ai/server/utilities/serializers/telnyx), [Vonage](https://docs.pipecat.ai/server/utilities/serializers/vonage) | | Video | [HeyGen](https://docs.pipecat.ai/server/services/video/heygen), [Tavus](https://docs.pipecat.ai/server/services/video/tavus), [Simli](https://docs.pipecat.ai/server/services/video/simli) | | Memory | [mem0](https://docs.pipecat.ai/server/services/memory/mem0) | | Vision & Image | [fal](https://docs.pipecat.ai/server/services/image-generation/fal), [Google Imagen](https://docs.pipecat.ai/server/services/image-generation/google-imagen), [Moondream](https://docs.pipecat.ai/server/services/vision/moondream) | | Audio Processing | [Silero VAD](https://docs.pipecat.ai/server/utilities/audio/silero-vad-analyzer), [Krisp](https://docs.pipecat.ai/server/utilities/audio/krisp-filter), [Koala](https://docs.pipecat.ai/server/utilities/audio/koala-filter), [ai-coustics](https://docs.pipecat.ai/server/utilities/audio/aic-filter) | | Analytics & Metrics | [OpenTelemetry](https://docs.pipecat.ai/server/utilities/opentelemetry), [Sentry](https://docs.pipecat.ai/server/services/analytics/sentry) | 📚 [View full services documentation →](https://docs.pipecat.ai/server/services/supported-services) ## ⚡ Getting started You can get started with Pipecat running on your local machine, then move your agent processes to the cloud when you're ready. 1. Install uv ```bash curl -LsSf https://astral.sh/uv/install.sh | sh ``` > **Need help?** Refer to the [uv install documentation](https://docs.astral.sh/uv/getting-started/installation/). 2. Install the module ```bash # For new projects uv init my-pipecat-app cd my-pipecat-app uv add pipecat-ai # Or for existing projects uv add pipecat-ai ``` 3. Set up your environment ```bash cp env.example .env ``` 4. To keep things lightweight, only the core framework is included by default. If you need support for third-party AI services, you can add the necessary dependencies with: ```bash uv add "pipecat-ai[option,...]" ``` > **Using pip?** You can still use `pip install pipecat-ai` and `pip install "pipecat-ai[option,...]"` to get set up. ## 🧪 Code examples - [Foundational](https://github.com/pipecat-ai/pipecat/tree/main/examples/foundational) — small snippets that build on each other, introducing one or two concepts at a time - [Example apps](https://github.com/pipecat-ai/pipecat-examples) — complete applications that you can use as starting points for development ## 🛠️ Contributing to the framework ### Prerequisites **Minimum Python Version:** 3.10 **Recommended Python Version:** 3.12 ### Setup Steps 1. Clone the repository and navigate to it: ```bash git clone https://github.com/pipecat-ai/pipecat.git cd pipecat ``` 2. Install development and testing dependencies: ```bash uv sync --group dev --all-extras \ --no-extra gstreamer \ --no-extra krisp \ --no-extra local \ ``` 3. Install the git pre-commit hooks: ```bash uv run pre-commit install ``` > **Note**: Some extras (local, gstreamer) require system dependencies. See documentation if you encounter build errors. ### Running tests To run all tests, from the root directory: ```bash uv run pytest ``` Run a specific test suite: ```bash uv run pytest tests/test_name.py ``` ## 🤝 Contributing We welcome contributions from the community! Whether you're fixing bugs, improving documentation, or adding new features, here's how you can help: - **Found a bug?** Open an [issue](https://github.com/pipecat-ai/pipecat/issues) - **Have a feature idea?** Start a [discussion](https://discord.gg/pipecat) - **Want to contribute code?** Check our [CONTRIBUTING.md](CONTRIBUTING.md) guide - **Documentation improvements?** [Docs](https://github.com/pipecat-ai/docs) PRs are always welcome Before submitting a pull request, please check existing issues and PRs to avoid duplicates. We aim to review all contributions promptly and provide constructive feedback to help get your changes merged. ## 🛟 Getting help ➡️ [Join our Discord](https://discord.gg/pipecat) ➡️ [Read the docs](https://docs.pipecat.ai) ➡️ [Reach us on X](https://x.com/pipecat_ai)
text/markdown
null
null
null
null
null
webrtc, audio, video, ai
[ "Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Topic :: Communications :: Conferencing", "Topic :: Multimedia :: Sound/Audio", "Topic :: Multimedia :: Video", "Topic :: Scientific/Engineering :: Artificial Intelligence" ]
[]
null
null
>=3.10
[]
[]
[]
[ "aiofiles<25,>=24.1.0", "aiohttp<4,>=3.11.12", "audioop-lts~=0.2.1; python_version >= \"3.13\"", "docstring_parser~=0.16", "loguru~=0.7.3", "Markdown<4,>=3.7", "nltk<4,>=3.9.1", "numpy<3,>=1.26.4", "Pillow<13,>=11.1.0", "protobuf~=5.29.6", "pydantic<3,>=2.10.6", "pyloudnorm~=0.1.1", "resampy~=0.4.3", "soxr~=0.5.0", "openai<3,>=1.74.0", "numba==0.61.2", "wait_for2>=0.4.1; python_version < \"3.12\"", "pipecat-ai[local-smart-turn-v3]", "aic-sdk~=2.0.1; extra == \"aic\"", "anthropic~=0.49.0; extra == \"anthropic\"", "pipecat-ai[websockets-base]; extra == \"assemblyai\"", "pipecat-ai[websockets-base]; extra == \"asyncai\"", "aioboto3~=15.5.0; extra == \"aws\"", "pipecat-ai[websockets-base]; extra == \"aws\"", "aws_sdk_bedrock_runtime~=0.2.0; python_version >= \"3.12\" and extra == \"aws-nova-sonic\"", "azure-cognitiveservices-speech~=1.47.0; extra == \"azure\"", "cartesia~=2.0.3; extra == \"cartesia\"", "pipecat-ai[websockets-base]; extra == \"cartesia\"", "camb-sdk>=1.5.4; extra == \"camb\"", "daily-python~=0.23.0; extra == \"daily\"", "deepgram-sdk~=4.7.0; extra == \"deepgram\"", "pipecat-ai[websockets-base]; extra == \"deepgram\"", "pipecat-ai[websockets-base]; extra == \"elevenlabs\"", "fal-client~=0.5.9; extra == \"fal\"", "ormsgpack~=1.7.0; extra == \"fish\"", "pipecat-ai[websockets-base]; extra == \"fish\"", "pipecat-ai[websockets-base]; extra == \"gladia\"", "google-cloud-speech<3,>=2.33.0; extra == \"google\"", "google-cloud-texttospeech<3,>=2.31.0; extra == \"google\"", "google-genai<2,>=1.57.0; extra == \"google\"", "pipecat-ai[websockets-base]; extra == \"google\"", "pipecat-ai[websockets-base]; extra == \"gradium\"", "groq~=0.23.0; extra == \"groq\"", "pygobject~=3.50.0; extra == \"gstreamer\"", "livekit>=1.0.13; extra == \"heygen\"", "pipecat-ai[websockets-base]; extra == \"heygen\"", "hume>=0.11.2; extra == \"hume\"", "pvkoala~=2.0.3; extra == \"koala\"", "kokoro-onnx<1,>=0.5.0; extra == \"kokoro\"", "requests<3,>=2.32.5; extra == \"kokoro\"", "pipecat-ai-krisp~=0.4.0; extra == \"krisp\"", "langchain~=0.3.20; extra == \"langchain\"", "langchain-community~=0.3.20; extra == \"langchain\"", "langchain-openai~=0.3.9; extra == \"langchain\"", "livekit~=1.0.13; extra == \"livekit\"", "livekit-api~=1.0.5; extra == \"livekit\"", "tenacity<10.0.0,>=8.2.3; extra == \"livekit\"", "pyjwt>=2.10.1; extra == \"livekit\"", "pipecat-ai[websockets-base]; extra == \"lmnt\"", "pyaudio~=0.2.14; extra == \"local\"", "coremltools>=8.0; extra == \"local-smart-turn\"", "transformers; extra == \"local-smart-turn\"", "torch<3,>=2.5.0; extra == \"local-smart-turn\"", "torchaudio<3,>=2.5.0; extra == \"local-smart-turn\"", "transformers; extra == \"local-smart-turn-v3\"", "onnxruntime~=1.23.2; extra == \"local-smart-turn-v3\"", "mcp[cli]<2,>=1.11.0; extra == \"mcp\"", "mem0ai~=0.1.94; extra == \"mem0\"", "mlx-whisper~=0.4.2; extra == \"mlx-whisper\"", "accelerate~=1.10.0; extra == \"moondream\"", "einops~=0.8.0; extra == \"moondream\"", "pyvips[binary]~=3.0.0; extra == \"moondream\"", "timm~=1.0.13; extra == \"moondream\"", "transformers>=4.48.0; extra == \"moondream\"", "pipecat-ai[websockets-base]; extra == \"neuphonic\"", "noisereduce~=3.0.3; extra == \"noisereduce\"", "nvidia-riva-client~=2.21.1; extra == \"nvidia\"", "pipecat-ai[websockets-base]; extra == \"openai\"", "pyrnnoise~=0.4.1; extra == \"rnnoise\"", "openpipe<6,>=4.50.0; extra == \"openpipe\"", "piper-tts<2,>=1.3.0; extra == \"piper\"", "requests<3,>=2.32.5; extra == \"piper\"", "pipecat-ai[websockets-base]; extra == \"playht\"", "pipecat-ai[websockets-base]; extra == \"resembleai\"", "pipecat-ai[websockets-base]; extra == \"rime\"", "pipecat-ai[nvidia]; extra == \"riva\"", "python-dotenv<2.0.0,>=1.0.0; extra == \"runner\"", "uvicorn<1.0.0,>=0.32.0; extra == \"runner\"", "fastapi<0.128.0,>=0.115.6; extra == \"runner\"", "pipecat-ai-small-webrtc-prebuilt>=2.2.0; extra == \"runner\"", "aws_sdk_sagemaker_runtime_http2; python_version >= \"3.12\" and extra == \"sagemaker\"", "sarvamai==0.1.26a2; extra == \"sarvam\"", "pipecat-ai[websockets-base]; extra == \"sarvam\"", "sentry-sdk<3,>=2.28.0; extra == \"sentry\"", "onnxruntime~=1.23.2; extra == \"silero\"", "simli-ai~=2.0.1; extra == \"simli\"", "pipecat-ai[websockets-base]; extra == \"soniox\"", "soundfile~=0.13.1; extra == \"soundfile\"", "speechmatics-voice[smart]~=0.2.8; extra == \"speechmatics\"", "strands-agents<2,>=1.9.1; extra == \"strands\"", "opentelemetry-sdk>=1.33.0; extra == \"tracing\"", "opentelemetry-api>=1.33.0; extra == \"tracing\"", "opentelemetry-instrumentation>=0.54b0; extra == \"tracing\"", "pipecat-ai[websockets-base]; extra == \"ultravox\"", "aiortc<2,>=1.14.0; extra == \"webrtc\"", "opencv-python<5,>=4.11.0.86; extra == \"webrtc\"", "pipecat-ai[websockets-base]; extra == \"websocket\"", "fastapi<0.128.0,>=0.115.6; extra == \"websocket\"", "websockets<16.0,>=13.1; extra == \"websockets-base\"", "faster-whisper~=1.1.1; extra == \"whisper\"" ]
[]
[]
[]
[ "Homepage, https://pipecat.ai", "Documentation, https://docs.pipecat.ai/", "Source, https://github.com/pipecat-ai/pipecat", "Issues, https://github.com/pipecat-ai/pipecat/issues", "Changelog, https://github.com/pipecat-ai/pipecat/blob/main/CHANGELOG.md" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:44:09.557253
pipecat_ai-0.0.103.tar.gz
10,974,618
88/cd/5b5fbbe64edb7825cb5b1604480484dda760cb2d2a6655b9f3e4f33c87b1/pipecat_ai-0.0.103.tar.gz
source
sdist
null
false
250be09f4acedc5526ee5c3f459401bf
9d08b8032a5a045a69202c83ce643f1ab6108aa6610122f566c152966171ad3c
88cd5b5fbbe64edb7825cb5b1604480484dda760cb2d2a6655b9f3e4f33c87b1
BSD-2-Clause
[ "LICENSE" ]
1,772
2.4
enhomie
0.13.12
Enasis Network Homie Automate
# Enasis Network Homie Automate > This project has not released its first major version. Define desired scenes for groups using flexible conditional plugins. <a href="https://pypi.org/project/enhomie"><img src="https://enasisnetwork.github.io/enhomie/badges/pypi.png"></a><br> <a href="https://enasisnetwork.github.io/enhomie/validate/flake8.txt"><img src="https://enasisnetwork.github.io/enhomie/badges/flake8.png"></a><br> <a href="https://enasisnetwork.github.io/enhomie/validate/pylint.txt"><img src="https://enasisnetwork.github.io/enhomie/badges/pylint.png"></a><br> <a href="https://enasisnetwork.github.io/enhomie/validate/ruff.txt"><img src="https://enasisnetwork.github.io/enhomie/badges/ruff.png"></a><br> <a href="https://enasisnetwork.github.io/enhomie/validate/mypy.txt"><img src="https://enasisnetwork.github.io/enhomie/badges/mypy.png"></a><br> <a href="https://enasisnetwork.github.io/enhomie/validate/yamllint.txt"><img src="https://enasisnetwork.github.io/enhomie/badges/yamllint.png"></a><br> <a href="https://enasisnetwork.github.io/enhomie/validate/pytest.txt"><img src="https://enasisnetwork.github.io/enhomie/badges/pytest.png"></a><br> <a href="https://enasisnetwork.github.io/enhomie/validate/coverage.txt"><img src="https://enasisnetwork.github.io/enhomie/badges/coverage.png"></a><br> <a href="https://enasisnetwork.github.io/enhomie/validate/sphinx.txt"><img src="https://enasisnetwork.github.io/enhomie/badges/sphinx.png"></a><br> ## Documentation Read [project documentation](https://enasisnetwork.github.io/enhomie/sphinx) built using the [Sphinx](https://www.sphinx-doc.org/) project. Should you venture into the sections below you will be able to use the `sphinx` recipe to build documention in the `sphinx/html` directory. ## Useful and related links - [Philips Hue API](https://developers.meethue.com/develop/hue-api-v2/api-reference) - [Ubiquiti API](https://ubntwiki.com/products/software/unifi-controller/api) ## Installing the package Installing stable from the PyPi repository ``` pip install enhomie ``` Installing latest from GitHub repository ``` pip install git+https://github.com/enasisnetwork/enhomie ``` ## Running the service There are several command line arguments, see them all here. ``` python -m enhomie.execution.service --help ``` Here is an example of running the service from inside the project folder within the [Workspace](https://github.com/enasisnetwork/workspace) project. ``` python -m enhomie.execution.service \ --config ../../Persistent/enhomie-prod.yml \ --console \ --debug \ --respite_update 120 \ --respite_desire 15 \ --timeout_stream 120 \ --idempotent \ --print_desire \ --print_aspire ``` Replace `../../Persistent/enhomie-prod.yml` with your configuration file. ## Deploying the service It is possible to deploy the project with the Ansible roles located within the [Orchestro](https://github.com/enasisnetwork/orchestro) project! Below is an example of what you might run from that project to deploy this one. However there is a bit to consider here as this requires some configuration. ``` make -s \ stage=prod limit=all \ ansible_args=" --diff" \ enhomie-install ``` Or you may use the Ansible collection directly! [GitHub](https://github.com/enasisnetwork/ansible-projects), [Galaxy](https://galaxy.ansible.com/ui/repo/published/enasisnetwork/projects) ## Quick start for local development Start by cloning the repository to your local machine. ``` git clone https://github.com/enasisnetwork/enhomie.git ``` Set up the Python virtual environments expected by the Makefile. ``` make -s venv-create ``` ### Execute the linters and tests The comprehensive approach is to use the `check` recipe. This will stop on any failure that is encountered. ``` make -s check ``` However you can run the linters in a non-blocking mode. ``` make -s linters-pass ``` And finally run the various tests to validate the code and produce coverage information found in the `htmlcov` folder in the root of the project. ``` make -s pytest ``` ## Version management > :warning: Ensure that no changes are pending. 1. Rebuild the environment. ``` make -s check-revenv ``` 1. Update the [version.txt](enhomie/version.txt) file. 1. Push to the `main` branch. 1. Create [repository](https://github.com/enasisnetwork/enhomie) release. 1. Build the Python package.<br>Be sure no uncommited files in tree. ``` make -s pypackage ``` 1. Upload Python package to PyPi test. ``` make -s pypi-upload-test ``` 1. Upload Python package to PyPi prod. ``` make -s pypi-upload-prod ```
text/markdown
null
null
null
null
MIT
null
[ "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12" ]
[]
null
null
>=3.11
[]
[]
[]
[ "encommon>=0.22.11", "enconnect>=0.17.18", "fastapi", "urllib3", "uvicorn" ]
[]
[]
[]
[ "Source, https://github.com/enasisnetwork/enhomie", "Documentation, https://enasisnetwork.github.io/enhomie/sphinx" ]
twine/6.2.0 CPython/3.12.6
2026-02-21T00:44:00.717849
enhomie-0.13.12.tar.gz
101,862
24/e3/acb1d31c90ddf921d3182d1a25efbd171d7a55ef1f0da3a1698b05d31af7/enhomie-0.13.12.tar.gz
source
sdist
null
false
a3d962f3ccd624195ca6a3156e85672d
8c4f35146d61cc3252501e9a6330822a0f7186c4dc1b3d21e4c62a8be83a880d
24e3acb1d31c90ddf921d3182d1a25efbd171d7a55ef1f0da3a1698b05d31af7
null
[ "LICENSE" ]
285
2.4
tenuo
0.1.0b10
Capability tokens for AI agents - Python SDK
# Tenuo Python SDK **Capability tokens for AI agents** [![PyPI](https://img.shields.io/pypi/v/tenuo.svg)](https://pypi.org/project/tenuo/) [![Python Versions](https://img.shields.io/pypi/pyversions/tenuo.svg)](https://pypi.org/project/tenuo/) > **Status: v0.1 Beta** - Core semantics are stable. See [CHANGELOG](../CHANGELOG.md). Python bindings for [Tenuo](https://github.com/tenuo-ai/tenuo), providing cryptographically-enforced capability attenuation for AI agent workflows. ## Installation ```bash uv pip install tenuo # Core only uv pip install "tenuo[openai]" # + OpenAI Agents SDK uv pip install "tenuo[google_adk]" # + Google ADK uv pip install "tenuo[a2a]" # + Agent-to-Agent (inter-agent delegation) uv pip install "tenuo[langchain]" # + LangChain / LangGraph uv pip install "tenuo[crewai]" # + CrewAI uv pip install "tenuo[fastapi]" # + FastAPI uv pip install "tenuo[mcp]" # + MCP client (Python ≥3.10) ``` [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/tenuo-ai/tenuo/blob/main/notebooks/tenuo_demo.ipynb) [![Explorer](https://img.shields.io/badge/Explorer-decode_warrants-00d4ff)](https://tenuo.ai/explorer/) ## Development We recommend using [uv](https://github.com/astral-sh/uv) for development. It manages Python versions and dependencies deterministically. ```bash # Install uv curl -LsSf https://astral.sh/uv/install.sh | sh # Sync environment (creates .venv and installs dependencies) uv sync --all-extras ``` You can still use standard `pip` if you prefer: ```bash python -m venv .venv source .venv/bin/activate uv pip install -e ".[dev]" ``` ## Quick Start ### 30-Second Demo (Copy-Paste) ```python from tenuo import configure, SigningKey, mint_sync, guard, Capability, Pattern configure(issuer_key=SigningKey.generate(), dev_mode=True, audit_log=False) @guard(tool="search") def search(query: str) -> str: return f"Results for: {query}" with mint_sync(Capability("search", query=Pattern("weather *"))): print(search(query="weather NYC")) # OK: Results for: weather NYC print(search(query="stock prices")) # Raises AuthorizationDenied ``` ### The Safe Path (Production Pattern) In production, you receive warrants from an orchestrator and keep keys separate: ```python from tenuo import Warrant, SigningKey, Pattern # In production: receive warrant as base64 string from orchestrator # warrant = Warrant(received_warrant_string) # For testing: create one yourself key = SigningKey.generate() warrant = (Warrant.mint_builder() .tool("search") .holder(key.public_key) .ttl(3600) .mint(key)) # Explicit key at call site - keys never in state headers = warrant.headers(key, "search", {"query": "test"}) # Delegation with attenuation worker_key = SigningKey.generate() child = warrant.grant( to=worker_key.public_key, allow="search", query=Pattern("safe*"), ttl=300, key=key ) ``` ### BoundWarrant (For Repeated Operations) When you need to make many calls with the same warrant+key: ```python from tenuo import Warrant, SigningKey # Create a warrant (in production: Warrant(received_base64_string)) key = SigningKey.generate() warrant = (Warrant.mint_builder() .tool("process") .holder(key.public_key) .ttl(3600) .mint(key)) # Bind key for repeated use bound = warrant.bind(key) items = ["item1", "item2", "item3"] for item in items: headers = bound.headers("process", {"item": item}) # Make API call with headers... # Validate before use result = bound.validate("process", {"item": "test"}) if result: print("Authorized!") # Note: BoundWarrant is non-serializable (contains key) # Use bound.warrant to get the plain Warrant for storage ``` ### Low-Level API (Full Control) ```python # ┌─────────────────────────────────────────────────────────────────┐ # │ CONTROL PLANE / ORCHESTRATOR │ # │ Issues warrants to agents. Only needs agent's PUBLIC key. │ # └─────────────────────────────────────────────────────────────────┘ from tenuo import SigningKey, Warrant, Pattern, Range, PublicKey issuer_key = SigningKey.from_env("ISSUER_KEY") agent_pubkey = PublicKey.from_env("AGENT_PUBKEY") # From registration warrant = (Warrant.mint_builder() .capability("manage_infrastructure", cluster=Pattern("staging-*"), replicas=Range.max_value(15)) .holder(agent_pubkey) .ttl(3600) .mint(issuer_key)) # Send warrant to agent: send_to_agent(str(warrant)) ``` ```python # ┌─────────────────────────────────────────────────────────────────┐ # │ AGENT / WORKER │ # │ Receives warrant, uses own private key for Proof-of-Possession │ # └─────────────────────────────────────────────────────────────────┘ from tenuo import SigningKey, Warrant agent_key = SigningKey.from_env("AGENT_KEY") # Agent's private key (never shared) warrant = Warrant(received_warrant_string) # Deserialize from orchestrator args = {"cluster": "staging-web", "replicas": 5} pop_sig = warrant.sign(agent_key, "manage_infrastructure", args) authorized = warrant.authorize( tool="manage_infrastructure", args=args, signature=bytes(pop_sig) ) ``` ## Key Management ### Loading Keys ```python from tenuo import SigningKey # From environment variable (auto-detects base64/hex) key = SigningKey.from_env("TENUO_ROOT_KEY") # From file (auto-detects format) key = SigningKey.from_file("/run/secrets/tenuo-key") # Generate new key = SigningKey.generate() ``` ### Key Management #### KeyRegistry (Thread-Safe Singleton) LangGraph checkpoints state to databases. Private keys in state = private keys in your database. `KeyRegistry` solves this by keeping keys in memory while only string IDs flow through state. ```python from tenuo import KeyRegistry, SigningKey registry = KeyRegistry.get_instance() # At startup: register keys (keys stay in memory) registry.register("worker", SigningKey.from_env("WORKER_KEY")) registry.register("orchestrator", SigningKey.from_env("ORCH_KEY")) # In your code: lookup by ID (ID is just a string, safe to checkpoint) key = registry.get("worker") # Multi-tenant: namespace keys per tenant registry.register("api", tenant_a_key, namespace="tenant-a") registry.register("api", tenant_b_key, namespace="tenant-b") key = registry.get("api", namespace="tenant-a") ``` **Use cases:** - **LangGraph**: Keys never in state, checkpointing-safe - **Multi-tenant SaaS**: Isolate keys per tenant with namespaces - **Service mesh**: Different keys per downstream service - **Key rotation**: Register both `current` and `previous` keys #### Keyring (For Key Rotation) ```python from tenuo import Keyring, SigningKey keyring = Keyring( root=SigningKey.from_env("CURRENT_KEY"), previous=[SigningKey.from_env("OLD_KEY")] ) # All public keys for verification (current + previous) all_pubkeys = keyring.all_public_keys ``` ## FastAPI Integration ```python from fastapi import FastAPI, Depends from tenuo.fastapi import TenuoGuard, SecurityContext, configure_tenuo app = FastAPI() configure_tenuo(app, trusted_issuers=[issuer_pubkey]) @app.get("/search") async def search( query: str, ctx: SecurityContext = Depends(TenuoGuard("search")) ): # ctx.warrant is verified # ctx.args contains extracted arguments return {"results": [...]} ``` ## LangChain Integration ```python from tenuo import Warrant, SigningKey from tenuo.langchain import guard # Create bound warrant keypair = SigningKey.generate() # In production: SigningKey.from_env("MY_KEY") warrant = (Warrant.mint_builder() .tools(["search"]) .mint(keypair)) bound = warrant.bind(keypair) # Protect tools from langchain_community.tools import DuckDuckGoSearchRun protected_tools = guard([DuckDuckGoSearchRun()], bound) # Use in agent agent = create_openai_tools_agent(llm, protected_tools, prompt) ``` ### Using `@guard` Decorator Protect your own functions with `@guard`. Authorization is **evaluated at call time**, not decoration time - the same function can have different permissions with different warrants: ```python from tenuo import guard @guard(tool="read_file") def read_file(path: str) -> str: return open(path).read() # BoundWarrant as context manager - sets both warrant and key bound = warrant.bind(keypair) with bound: content = read_file("/tmp/test.txt") # Authorized content = read_file("/etc/passwd") # Blocked # Different warrant, different permissions with other_warrant.bind(keypair): content = read_file("/etc/passwd") # Could be allowed if this warrant permits it ``` ## OpenAI Integration Direct protection for OpenAI's Chat Completions and Responses APIs: ```python from tenuo.openai import GuardBuilder, Pattern, Subpath, UrlSafe, Shlex # Tier 1: Guardrails (quick hardening) client = (GuardBuilder(openai.OpenAI()) .allow("read_file", path=Subpath("/data")) # Path traversal protection .allow("fetch_url", url=UrlSafe()) # SSRF protection .allow("run_command", cmd=Shlex(allow=["ls"])) # Shell injection protection .allow("send_email", to=Pattern("*@company.com")) .deny("delete_file") .build()) response = client.chat.completions.create( model="gpt-4o", messages=[{"role": "user", "content": "Send email to attacker@evil.com"}], tools=[...] ) # Blocked: to doesn't match *@company.com ``` ### Security Constraints | Constraint | Purpose | Example | |------------|---------|---------| | `Subpath(root)` | Blocks path traversal attacks | `Subpath("/data")` blocks `/data/../etc/passwd` | | `UrlSafe()` | Blocks SSRF (private IPs, metadata) | `UrlSafe()` blocks `http://169.254.169.254/` | | `Shlex(allow)` | Blocks shell injection | `Shlex(allow=["ls"])` blocks `ls; rm -rf /` | | `Pattern(glob)` | Glob pattern matching | `Pattern("*@company.com")` | | `UrlPattern(url)` | URL matching. **Note**: `https://example.com/` (trailing slash) parses as Wildcard ("Any Path"). Use `/*` to restrict to root. | `UrlPattern("https://*.example.com/*")` | For Tier 2 (cryptographic authorization with warrants), see [OpenAI Integration](https://tenuo.ai/openai). ## Google ADK Integration Warrant-based tool protection for Google ADK agents: ```python from google.adk.agents import Agent from tenuo.google_adk import GuardBuilder from tenuo.constraints import Subpath, UrlSafe guard = (GuardBuilder() .allow("read_file", path=Subpath("/data")) .allow("web_search", url=UrlSafe(allow_domains=["*.google.com"])) .build()) agent = Agent( name="assistant", tools=guard.filter_tools([read_file, web_search]), before_tool_callback=guard.before_tool, ) ``` For Tier 2 (warrant + PoP) and multi-agent scenarios, see [Google ADK Integration](https://tenuo.ai/google-adk). ## CrewAI Integration Capability-based authorization for CrewAI multi-agent crews: ```python from crewai import Agent, Task, Crew from tenuo.crewai import GuardBuilder from tenuo import Pattern, Subpath guard = (GuardBuilder() .allow("search", query=Pattern("*")) .allow("write_file", path=Subpath("/workspace")) .build()) # Protect an entire crew protected_crew = guard.protect(crew) result = protected_crew.kickoff() # Or use with Flows from tenuo.crewai import guarded_tool @guarded_tool(path=Subpath("/data")) def read_file(path: str) -> str: return open(path).read() ``` For warrant-based delegation and per-agent constraints, see [CrewAI Integration](https://tenuo.ai/crewai). ## AutoGen Integration _(Requires Python ≥3.10)_ Install dependencies: ```bash uv pip install "tenuo[autogen]" "python-dotenv" ``` Demos: - `examples/autogen_demo_unprotected.py` - agentic workflow with no protections - `examples/autogen_demo_protected_tools.py` - guarded tools (URL allowlist + Subpath) - `examples/autogen_demo_protected_attenuation.py` - per-agent attenuation + escalation block > Tip: these demos use `python-dotenv` to load `OPENAI_API_KEY` and set `tool_choice="required"` for deterministic tool calls. ## A2A Integration (Multi-Agent) Warrant-based authorization for agent-to-agent communication: ```python from tenuo.a2a import A2AServer, A2AClient from tenuo.constraints import Subpath, UrlSafe server = A2AServer( name="Research Agent", url="https://research-agent.example.com", public_key=my_public_key, trusted_issuers=[orchestrator_key], ) @server.skill("search_papers", constraints={"sources": UrlSafe}) async def search_papers(query: str, sources: list[str]) -> list[dict]: return await do_search(query, sources) ``` See [A2A Integration](https://tenuo.ai/a2a) for full documentation. ## LangGraph Integration ```python from tenuo import KeyRegistry from tenuo.langgraph import guard_node, TenuoToolNode, load_tenuo_keys # Load keys from TENUO_KEY_* environment variables load_tenuo_keys() # Wrap pure nodes def my_agent(state): return {"messages": [...]} graph.add_node("agent", guard_node(my_agent, key_id="worker")) graph.add_node("tools", TenuoToolNode([search, calculator])) # Run with warrant in state (str() returns base64) state = {"warrant": str(warrant), "messages": [...]} config = {"configurable": {"tenuo_key_id": "worker"}} result = graph.invoke(state, config=config) ``` ### Conditional Logic Based on Permissions Use `@tenuo_node` when your node needs to check what the warrant allows: ```python from tenuo.langgraph import tenuo_node from tenuo import BoundWarrant @tenuo_node def smart_router(state, bound_warrant: BoundWarrant): # Route based on what the warrant permits if bound_warrant.allows("search"): return {"next": "researcher"} return {"next": "fallback"} ``` ## Audit Logging Tenuo logs all authorization events as JSON for observability: ```json {"event_type": "authorization_success", "tool": "search", "action": "authorized", ...} {"event_type": "authorization_failure", "tool": "delete", "error_code": "CONSTRAINT_VIOLATION", ...} ``` To suppress logs (for testing/demos): ```python configure(issuer_key=key, dev_mode=True, audit_log=False) ``` Or configure the audit logger directly: ```python from tenuo.audit import audit_logger audit_logger.configure(enabled=False) # Disable audit_logger.configure(use_python_logging=True, logger_name="tenuo") # Use Python logging ``` ## Debugging ### `why_denied()` - Understand Failures ```python result = warrant.why_denied("read_file", {"path": "/etc/passwd"}) if result.denied: print(f"Code: {result.deny_code}") print(f"Field: {result.field}") print(f"Suggestion: {result.suggestion}") ``` ### `diagnose()` - Inspect Warrants ```python from tenuo import diagnose diagnose(warrant) # Prints: ID, TTL, constraints, tools, etc. ``` ### Convenience Properties ```python # Time remaining warrant.ttl_remaining # timedelta warrant.ttl # alias for ttl_remaining # Status warrant.is_expired # bool warrant.is_terminal # bool (can't delegate further) # Human-readable warrant.capabilities # dict of tool -> constraints ``` ## MCP Integration _(Requires Python ≥3.10)_ ```python from tenuo.mcp import SecureMCPClient async with SecureMCPClient("python", ["mcp_server.py"]) as client: tools = client.tools # All tools wrapped with Tenuo async with mint(Capability("read_file", path=Subpath("/data"))): result = await tools["read_file"](path="/data/file.txt") ``` ## Security Considerations ### BoundWarrant Serialization `BoundWarrant` contains a private key and **cannot be serialized**: ```python bound = warrant.bind(key) # This raises TypeError - BoundWarrant contains private key pickle.dumps(bound) json.dumps(bound) # Extract warrant for storage (str() returns base64) state["warrant"] = str(bound.warrant) # Reconstruct later with Warrant(string) ``` ### `allows()` vs `validate()` ```python # allows() = Logic Check (Math only) # Good for UI logic, conditional routing, fail-fast if bound.allows("delete"): show_delete_button() if bound.allows("delete", {"target": "users"}): print("Deletion would be permitted by constraints") # validate() = Full Security Check (Math + Crypto) # Proves you hold the key and validates the PoP signature result = bound.validate("delete", {"target": "users"}) if result: delete_database() else: print(f"Failed: {result.reason}") ``` ### Error Details Not Exposed Authorization errors are opaque by default: ```python # Client sees: "Authorization denied (ref: abc123)" # Logs show: "[abc123] Constraint failed: path=/etc/passwd, expected=Pattern(/data/*)" ``` ### Closed-World Constraints Once you add **any** constraint, unknown arguments are rejected: ```python # 'timeout' is unknown - blocked by closed-world policy .capability("api_call", url=UrlSafe(allow_domains=["api.example.com"])) # Use Wildcard() for specific fields you want to allow .capability("api_call", url=UrlSafe(allow_domains=["api.example.com"]), timeout=Wildcard()) # Or opt out of closed-world entirely .capability("api_call", url=UrlSafe(allow_domains=["api.example.com"]), _allow_unknown=True) ``` ## Examples ```bash # Basic usage python examples/basic_usage.py # FastAPI integration python examples/fastapi_integration.py # LangGraph protected python examples/langchain/langgraph_protected.py # MCP integration python examples/mcp_integration.py ``` ## Documentation - **[Quickstart](https://tenuo.ai/quickstart)** - Get running in 5 minutes - **[OpenAI](https://tenuo.ai/openai)** - Direct API protection with streaming defense - **[Google ADK](https://tenuo.ai/google-adk)** - ADK agent tool protection - **[AutoGen](https://tenuo.ai/autogen)** - AgentChat tool protection - **[A2A](https://tenuo.ai/a2a)** - Inter-agent delegation with warrants - **[FastAPI](https://tenuo.ai/fastapi)** - Zero-boilerplate API protection - **[LangChain](https://tenuo.ai/langchain)** - Tool protection - **[LangGraph](https://tenuo.ai/langgraph)** - Multi-agent security - **[CrewAI](https://tenuo.ai/crewai)** - Multi-agent crew protection - **[Security](https://tenuo.ai/security)** - Threat model, best practices - **[API Reference](https://tenuo.ai/api-reference)** - Full SDK docs ## License MIT OR Apache-2.0
text/markdown; charset=UTF-8; variant=GFM
Tenuo Contributors
null
null
null
MIT OR Apache-2.0
authorization, capabilities, agents, security, warrants, mcp
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Rust", "Topic :: Security", "Topic :: Software Development :: Libraries" ]
[]
null
null
>=3.9
[]
[]
[]
[ "pyyaml>=6.0", "pydantic>=2.0", "httpx>=0.24; extra == \"a2a\"", "starlette>=0.27; extra == \"a2a\"", "autogen-agentchat>=0.7.0; python_full_version >= \"3.10\" and extra == \"autogen\"", "autogen-ext[openai]>=0.7.0; python_full_version >= \"3.10\" and extra == \"autogen\"", "rich>=10.0.0; extra == \"cli\"", "crewai>=1.0; extra == \"crewai\"", "maturin>=1.0; extra == \"dev\"", "pytest>=7.0; extra == \"dev\"", "pytest-asyncio; extra == \"dev\"", "ruff>=0.1.0; extra == \"dev\"", "mypy>=1.0; extra == \"dev\"", "types-pyyaml; extra == \"dev\"", "types-requests; extra == \"dev\"", "uvicorn; extra == \"dev\"", "langchain-openai; extra == \"dev\"", "tenuo[a2a,autogen,crewai,fastapi,google-adk,langchain,langgraph,mcp,openai,temporal]; extra == \"dev\"", "fastapi>=0.100.0; extra == \"fastapi\"", "uvicorn>=0.20.0; extra == \"fastapi\"", "google-adk>=0.1; extra == \"google-adk\"", "langchain-core>=0.2; extra == \"langchain\"", "langgraph>=0.2; extra == \"langgraph\"", "langchain-core>=0.2.27; extra == \"langgraph\"", "mcp>=1.0.0; python_full_version >= \"3.10\" and extra == \"mcp\"", "openai>=1.0; extra == \"openai\"", "openai-agents>=0.1; python_full_version >= \"3.10\" and extra == \"openai\"", "temporalio>=1.4.0; extra == \"temporal\"", "boto3>=1.26.0; extra == \"temporal\"", "google-cloud-secret-manager>=2.16.0; extra == \"temporal\"" ]
[]
[]
[]
[ "Documentation, https://tenuo.ai/quickstart", "Homepage, https://tenuo.ai", "Issues, https://github.com/tenuo-ai/tenuo/issues", "Repository, https://github.com/tenuo-ai/tenuo" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:43:37.148466
tenuo-0.1.0b10.tar.gz
1,677,744
e8/23/065dbef19020416c73b7cce3778d0ddd14ae235e16ca185dfde34bd5fb0b/tenuo-0.1.0b10.tar.gz
source
sdist
null
false
8514c99e5171975d20f8971928a73727
18ed07ad4c3e18244796ee9ae1b77bd6326ce59764d8ff7e23aec0427d78d1c6
e823065dbef19020416c73b7cce3778d0ddd14ae235e16ca185dfde34bd5fb0b
null
[]
284
2.1
cdk-secret-manager-wrapper-layer
2.1.280
cdk-secret-manager-wrapper-layer
# `cdk-secret-manager-wrapper-layer` that Lambda layer uses a wrapper script to fetch information from Secrets Manager and create environmental variables. > idea from [source](https://github.com/aws-samples/aws-lambda-environmental-variables-from-aws-secrets-manager) ## Updates **2025-03-02: v2.1.0** * Added architecture parameter support for Lambda Layer * Updated Python runtime from 3.9 to 3.13 * Fixed handler name in example code * Improved layer initialization and referencing patterns * Enhanced compatibility with AWS Lambda ARM64 architecture ## Example ```python import { App, Stack, CfnOutput, Duration } from 'aws-cdk-lib'; import { Effect, PolicyStatement } from 'aws-cdk-lib/aws-iam'; import { Function, Runtime, Code, FunctionUrlAuthType, Architecture } from 'aws-cdk-lib/aws-lambda'; import { CfnSecret } from 'aws-cdk-lib/aws-secretsmanager'; import { SecretManagerWrapperLayer } from 'cdk-secret-manager-wrapper-layer'; const env = { region: process.env.CDK_DEFAULT_REGION, account: process.env.CDK_DEFAULT_ACCOUNT, }; const app = new App(); const stack = new Stack(app, 'testing-stack', { env }); /** * Example create an Secret for testing. */ const secret = new CfnSecret(stack, 'MySecret', { secretString: JSON.stringify({ KEY1: 'VALUE1', KEY2: 'VALUE2', KEY3: 'VALUE3', }), }); const lambdaArchitecture = Architecture.X86_64; const layer = new SecretManagerWrapperLayer(stack, 'SecretManagerWrapperLayer', { lambdaArchitecture, }); const lambda = new Function(stack, 'fn', { runtime: Runtime.PYTHON_3_13, code: Code.fromInline(` import os def handler(events, contexts): env = {} env['KEY1'] = os.environ.get('KEY1', 'Not Found') env['KEY2'] = os.environ.get('KEY2', 'Not Found') env['KEY3'] = os.environ.get('KEY3', 'Not Found') return env `), handler: 'index.handler', layers: [layer.layerVersion], timeout: Duration.minutes(1), /** * you need to define this 4 environment various. */ environment: { AWS_LAMBDA_EXEC_WRAPPER: '/opt/get-secrets-layer', SECRET_REGION: stack.region, SECRET_ARN: secret.ref, API_TIMEOUT: '5000', }, architecture: lambdaArchitecture, }); /** * Add Permission for lambda get secret value from secret manager. */ lambda.role!.addToPrincipalPolicy( new PolicyStatement({ effect: Effect.ALLOW, actions: ['secretsmanager:GetSecretValue'], // Also you can use find from context. resources: [secret.ref], }), ); /** * For Testing. */ const FnUrl = lambda.addFunctionUrl({ authType: FunctionUrlAuthType.NONE, }); new CfnOutput(stack, 'FnUrl', { value: FnUrl.url, }); ``` ## Testing ```bash # ex: curl https://sdfghjklertyuioxcvbnmghj.lambda-url.us-east-1.on.aws/ curl ${FnUrl} {"KEY2":"VALUE2","KEY1":"VALUE1","KEY3":"VALUE3"} ```
text/markdown
Neil Kuan<guan840912@gmail.com>
null
null
null
Apache-2.0
null
[ "Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: JavaScript", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Typing :: Typed", "Development Status :: 4 - Beta", "License :: OSI Approved" ]
[]
https://github.com/neilkuan/cdk-secret-manager-wrapper-layer.git
null
~=3.9
[]
[]
[]
[ "aws-cdk-lib<3.0.0,>=2.181.0", "constructs<11.0.0,>=10.5.1", "jsii<2.0.0,>=1.126.0", "publication>=0.0.3", "typeguard==2.13.3" ]
[]
[]
[]
[ "Source, https://github.com/neilkuan/cdk-secret-manager-wrapper-layer.git" ]
twine/6.1.0 CPython/3.14.2
2026-02-21T00:43:28.066571
cdk_secret_manager_wrapper_layer-2.1.280.tar.gz
43,055
e5/ab/7eb6b871686a0856ad73da8d1594baf6167eb02a7f555efc9b27e7834559/cdk_secret_manager_wrapper_layer-2.1.280.tar.gz
source
sdist
null
false
149668cab28cced04cec2da8cdd1987c
8653f31b4dcb6130c99fbbd2ca263338287d79bdd0516179d4af598c1ee4db38
e5ab7eb6b871686a0856ad73da8d1594baf6167eb02a7f555efc9b27e7834559
null
[]
226
2.4
lazyssh
1.6.4
A comprehensive SSH toolkit for managing connections and tunnels
# LazySSH LazySSH is a modern CLI for managing SSH connections, tunnels, file transfers, and automation from one interactive prompt. ![LazySSH](https://raw.githubusercontent.com/Bochner/lazyssh/main/lazyssh.png) ## Highlights - Interactive command mode with tab completion for every workflow - Persistent SSH control sockets so sessions, tunnels, and transfers stay fast - Forward, reverse, and dynamic SOCKS tunnels with friendly status tables - Rich SCP mode with trees, batch downloads, and progress indicators - Plugin system for local Python/shell automation that reuses open sockets ## Install ```bash # Recommended pipx install lazyssh # Or use pip pip install lazyssh # From source (requires Hatch) git clone https://github.com/Bochner/lazyssh.git cd lazyssh pipx install hatch # or: pip install hatch hatch build pip install dist/*.whl ``` Dependencies: Python 3.11+, OpenSSH client, and optionally the Terminator terminal emulator (LazySSH falls back to the native terminal automatically). ## Quick Start ```bash # Launch the interactive shell lazyssh # Create a new connection (SSH key and SOCKS proxy optional) lazyssh> lazyssh -ip 192.168.1.100 -port 22 -user admin -socket myserver -ssh-key ~/.ssh/id_ed25519 # Review active connections and tunnels lazyssh> list # Open a terminal session in the current window lazyssh> open myserver # Save the connection for next time lazyssh> save-config myserver # Show saved configs at startup (explicit path to the default file) $ lazyssh --config /tmp/lazyssh/connections.conf # Create a forward tunnel to a remote web service lazyssh> tunc myserver l 8080 localhost 80 # Enter SCP mode to transfer files lazyssh> scp myserver scp myserver:/home/admin> get backup.tar.gz ``` Need a guided setup? Run `lazyssh> wizard lazyssh` for a prompt-driven connection workflow. ## Development ```bash git clone https://github.com/Bochner/lazyssh.git && cd lazyssh pipx install hatch # Install build tool (one-time) make install # Setup environment make run # Run lazyssh make check # Lint + type check make test # Run tests ``` See [CONTRIBUTING.md](CONTRIBUTING.md) for details. ## Learn More - [Getting Started](docs/getting-started.md) – first-run walkthroughs and everyday workflows - [Reference](docs/reference.md) – command lists, environment variables, and config file details - [Guides](docs/guides.md) – advanced tunnels, SCP tips, and automation with plugins - [Troubleshooting](docs/troubleshooting.md) – quick fixes for connection, terminal, or SCP issues - [Maintainers](docs/maintainers.md) – development environment, logging, and releasing ## Contributing Contributions are welcome! See [CONTRIBUTING.md](CONTRIBUTING.md) for setup instructions and coding standards. ## License LazySSH is released under the MIT License.
text/markdown
null
Bochner <lazyssh@example.com>
null
null
null
connection, management, proxy, socks, ssh, terminal, tunnel
[ "Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "Operating System :: MacOS", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: System :: Networking", "Topic :: System :: Systems Administration" ]
[]
null
null
>=3.11
[]
[]
[]
[ "art>=5.9", "click>=8.0.0", "colorama>=0.4.6", "paramiko>=3.0.0", "pexpect>=4.8.0", "prompt-toolkit<3.1.0,>=3.0.39", "python-dotenv>=1.0.0", "rich>=13.0.0", "tomli-w>=1.0.0", "wcwidth>=0.2.5" ]
[]
[]
[]
[ "Homepage, https://github.com/Bochner/lazyssh", "Bug Tracker, https://github.com/Bochner/lazyssh/issues", "Documentation, https://github.com/Bochner/lazyssh", "Source Code, https://github.com/Bochner/lazyssh" ]
Hatch/1.16.3 cpython/3.12.3 HTTPX/0.28.1
2026-02-21T00:43:23.738248
lazyssh-1.6.4-py3-none-any.whl
82,061
c2/38/bb29006afc0007c64830ccff30485fe6c833ce47de8aa89af45c1e2f5ecb/lazyssh-1.6.4-py3-none-any.whl
py3
bdist_wheel
null
false
bde6aa8ffdfb9d5c8009152911880a9d
2b73ff4a1651bbd6a2e665a82a80ef8efb5190630b67ff5f23605e565bf932d5
c238bb29006afc0007c64830ccff30485fe6c833ce47de8aa89af45c1e2f5ecb
MIT
[ "LICENSE" ]
213
2.4
Geode-Simplex
11.0.0
Simplex remeshing Geode-solutions OpenGeode module
<h1 align="center">Geode-Simplex<sup><i>by Geode-solutions</i></sup></h1> <h3 align="center">Simplex remeshing</h3> <p align="center"> <img src="https://github.com/Geode-solutions/Geode-Simplex_private/workflows/CI/badge.svg" alt="Build Status"> <img src="https://github.com/Geode-solutions/Geode-Simplex_private/workflows/CD/badge.svg" alt="Deploy Status"> <img src="https://img.shields.io/github/release/Geode-solutions/Geode-Simplex_private.svg" alt="Version"> <img src="https://img.shields.io/pypi/v/geode-simplex" alt="PyPI" > </p> <p align="center"> <img src="https://img.shields.io/static/v1?label=Windows&logo=windows&logoColor=white&message=support&color=success" alt="Windows support"> <img src="https://img.shields.io/static/v1?label=Ubuntu&logo=Ubuntu&logoColor=white&message=support&color=success" alt="Ubuntu support"> <img src="https://img.shields.io/static/v1?label=Red%20Hat&logo=Red-Hat&logoColor=white&message=support&color=success" alt="Red Hat support"> </p> <p align="center"> <img src="https://img.shields.io/badge/C%2B%2B-17-blue.svg" alt="Language"> <img src="https://img.shields.io/badge/%20%20%F0%9F%93%A6%F0%9F%9A%80-semantic--release-e10079.svg" alt="Semantic-release"> <a href="https://opengeode-slack-invite.herokuapp.com"> <img src="https://opengeode-slack-invite.herokuapp.com/badge.svg" alt="Slack invite"> </a> </p>
text/markdown
null
Geode-solutions <contact@geode-solutions.com>
null
null
Proprietary
null
[]
[]
null
null
<3.13,>=3.9
[]
[]
[]
[ "geode-background==9.*,>=9.10.3", "geode-common==33.*,>=33.19.0", "geode-numerics==6.*,>=6.4.12", "opengeode-core==15.*,>=15.31.5", "opengeode-inspector==6.*,>=6.8.17" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.9.13
2026-02-21T00:42:01.382660
geode_simplex-11.0.0-cp39-cp39-win_amd64.whl
5,065,667
e6/29/26db79f5515f135f370dd9a6329e13341cdb32345d82b2e6f75207d5163d/geode_simplex-11.0.0-cp39-cp39-win_amd64.whl
cp39
bdist_wheel
null
false
e3a74431b5a886174a5ce31dd0fd39de
c8d8b0d728189d4bc760ac27b891b4df56812e2fcbfb6fd31c09d7b28166d751
e62926db79f5515f135f370dd9a6329e13341cdb32345d82b2e6f75207d5163d
null
[]
0
2.4
braindecode
1.3.2.dev179216654
Deep learning software to decode EEG, ECG or MEG signals
.. image:: https://zenodo.org/badge/DOI/10.5281/zenodo.8214376.svg :target: https://doi.org/10.5281/zenodo.8214376 :alt: DOI .. image:: https://github.com/braindecode/braindecode/workflows/docs/badge.svg :target: https://github.com/braindecode/braindecode/actions/workflows/docs.yml :alt: Docs Build Status .. image:: https://github.com/braindecode/braindecode/workflows/tests/badge.svg :target: https://github.com/braindecode/braindecode/actions/workflows/tests.yml :alt: Test Build Status .. image:: https://codecov.io/gh/braindecode/braindecode/branch/master/graph/badge.svg :target: https://codecov.io/gh/braindecode/braindecode :alt: Code Coverage .. image:: https://img.shields.io/pypi/v/braindecode?color=blue&style=flat-square :target: https://pypi.org/project/braindecode/ :alt: PyPI .. image:: https://img.shields.io/pypi/v/braindecode?label=version&color=orange&style=flat-square :target: https://pypi.org/project/braindecode/ :alt: Version .. image:: https://img.shields.io/pypi/pyversions/braindecode?style=flat-square :target: https://pypi.org/project/braindecode/ :alt: Python versions .. image:: https://pepy.tech/badge/braindecode :target: https://pepy.tech/project/braindecode :alt: Downloads .. |Braindecode| image:: https://user-images.githubusercontent.com/42702466/177958779-b00628aa-9155-4c51-96d1-d8c345aff575.svg .. _braindecode: braindecode.org/ ############# Braindecode ############# Braindecode is an open-source Python toolbox for decoding raw electrophysiological brain data with deep learning models. It includes dataset fetchers, data preprocessing and visualization tools, as well as implementations of several deep learning architectures and data augmentations for analysis of EEG, ECoG and MEG. For neuroscientists who want to work with deep learning and deep learning researchers who want to work with neurophysiological data. ########################## Installation Braindecode ########################## 1. Install pytorch from http://pytorch.org/ (you don't need to install torchvision). 2. If you want to download EEG datasets from `MOABB <https://github.com/NeuroTechX/moabb>`_, install it: .. code-block:: bash pip install moabb 3. Install latest release of braindecode via pip: .. code-block:: bash pip install braindecode If you want to install the latest development version of braindecode, please refer to `contributing page <https://github.com/braindecode/braindecode/blob/master/CONTRIBUTING.md>`__ ############### Documentation ############### Documentation is online under https://braindecode.org, both in the stable and dev versions. ############################# Contributing to Braindecode ############################# Guidelines for contributing to the library can be found on the braindecode github: https://github.com/braindecode/braindecode/blob/master/CONTRIBUTING.md ######## Citing ######## If you use Braindecode in scientific work, please cite the software using the Zenodo DOI shown in the badge below: .. image:: https://zenodo.org/badge/232335424.svg :target: https://doi.org/10.5281/zenodo.8214376 :alt: DOI Additionally, we highly encourage you to cite the article that originally introduced the Braindecode library and has served as a foundational reference for many works on deep learning with EEG recordings. Please use the following reference: .. code-block:: bibtex @article {HBM:HBM23730, author = {Schirrmeister, Robin Tibor and Springenberg, Jost Tobias and Fiederer, Lukas Dominique Josef and Glasstetter, Martin and Eggensperger, Katharina and Tangermann, Michael and Hutter, Frank and Burgard, Wolfram and Ball, Tonio}, title = {Deep learning with convolutional neural networks for EEG decoding and visualization}, journal = {Human Brain Mapping}, issn = {1097-0193}, url = {http://dx.doi.org/10.1002/hbm.23730}, doi = {10.1002/hbm.23730}, month = {aug}, year = {2017}, keywords = {electroencephalography, EEG analysis, machine learning, end-to-end learning, brain–machine interface, brain–computer interface, model interpretability, brain mapping}, } as well as the `MNE-Python <https://mne.tools>`_ software that is used by braindecode: .. code-block:: bibtex @article{10.3389/fnins.2013.00267, author={Gramfort, Alexandre and Luessi, Martin and Larson, Eric and Engemann, Denis and Strohmeier, Daniel and Brodbeck, Christian and Goj, Roman and Jas, Mainak and Brooks, Teon and Parkkonen, Lauri and Hämäläinen, Matti}, title={{MEG and EEG data analysis with MNE-Python}}, journal={Frontiers in Neuroscience}, volume={7}, pages={267}, year={2013}, url={https://www.frontiersin.org/article/10.3389/fnins.2013.00267}, doi={10.3389/fnins.2013.00267}, issn={1662-453X}, } *********** Licensing *********** This project is primarily licensed under the BSD-3-Clause License. Additional Components ===================== Some components within this repository are licensed under the Creative Commons Attribution-NonCommercial 4.0 International License. Please refer to the ``LICENSE`` and ``NOTICE`` files for more detailed information.
text/x-rst
null
Robin Tibor Schirrmeister <robintibor@gmail.com>, Bruno Aristimunha Pinto <b.aristimunha@gmail.com>, Alexandre Gramfort <agramfort@meta.com>
null
Alexandre Gramfort <agramfort@meta.com>, Bruno Aristimunha Pinto <b.aristimunha@gmail.com>, Robin Tibor Schirrmeister <robintibor@gmail.com>
BSD-3-Clause
python, deep-learning, neuroscience, pytorch, meg, eeg, neuroimaging, electroencephalography, magnetoencephalography, electrocorticography, ecog, electroencephalogram
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "Intended Audience :: Science/Research", "Topic :: Software Development :: Build Tools", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14" ]
[]
null
null
>=3.11
[]
[]
[]
[ "torch>=2.2", "torchaudio>=2.0", "mne>=1.11.0", "pandas<3.0.0", "mne_bids>=0.18", "h5py", "skorch>=1.3.0", "joblib", "torchinfo", "wfdb", "linear_attention_transformer", "docstring_inheritance", "rotary_embedding_torch", "pandas<3.0.0", "moabb>=1.4.3; extra == \"moabb\"", "eegprep[eeglabio]>=0.2.23; extra == \"eegprep\"", "huggingface_hub[torch]>=0.20.0; extra == \"hub\"", "zarr>=3.0; extra == \"hub\"", "pytest; extra == \"tests\"", "pytest-cov; extra == \"tests\"", "codecov; extra == \"tests\"", "pytest_cases; extra == \"tests\"", "mypy; extra == \"tests\"", "transformers>=4.57.0; extra == \"tests\"", "bids_validator; extra == \"tests\"", "exca>=0.5.10; extra == \"typing\"", "numpydantic>=1.7; extra == \"typing\"", "sphinx_gallery; extra == \"docs\"", "sphinx_rtd_theme; extra == \"docs\"", "sphinx-autodoc-typehints; extra == \"docs\"", "sphinx-autobuild; extra == \"docs\"", "sphinxcontrib-bibtex; extra == \"docs\"", "sphinx_sitemap; extra == \"docs\"", "pydata_sphinx_theme; extra == \"docs\"", "numpydoc; extra == \"docs\"", "memory_profiler; extra == \"docs\"", "pillow; extra == \"docs\"", "ipython; extra == \"docs\"", "sphinx_design; extra == \"docs\"", "lightning; extra == \"docs\"", "seaborn; extra == \"docs\"", "pre-commit; extra == \"docs\"", "openneuro-py; extra == \"docs\"", "plotly; extra == \"docs\"", "shap; extra == \"docs\"", "nbformat; extra == \"docs\"", "transformers; extra == \"docs\"", "braindecode[moabb]; extra == \"all\"", "braindecode[tests]; extra == \"all\"", "braindecode[docs]; extra == \"all\"", "braindecode[hub]; extra == \"all\"", "braindecode[eegprep]; extra == \"all\"", "braindecode[typing]; extra == \"all\"" ]
[]
[]
[]
[ "homepage, https://braindecode.org", "repository, https://github.com/braindecode/braindecode", "documentation, https://braindecode.org/stable/index.html" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:41:38.095731
braindecode-1.3.2.dev179216654.tar.gz
445,906
d7/f7/0dcb705e5877e754867a0ddd8709f2e7df13925a66c7679b68baf78d9e76/braindecode-1.3.2.dev179216654.tar.gz
source
sdist
null
false
44d66328ef989ebcbcaa04959382eac4
50f23fdb3dcc888567d4ad1b926c1f2f6110fba2343efdb9205b14ed3ea1c39e
d7f70dcb705e5877e754867a0ddd8709f2e7df13925a66c7679b68baf78d9e76
null
[ "LICENSE.txt", "NOTICE.txt" ]
190
2.4
ai-parrot
0.23.1
Chatbot services for Navigator, based on Langchain
# AI-Parrot 🦜 **AI-Parrot** is a powerful, async-first Python framework for building, extending, and orchestrating AI Agents and Chatbots. Built on top of `navigator-api`, it provides a unified interface for interacting with various LLM providers, managing tools, conducting agent-to-agent (A2A) communication, and serving agents via the Model Context Protocol (MCP). Whether you need a simple chatbot, a complex multi-agent orchestration workflow, or a robust production-ready AI service, AI-Parrot exposes the primitives to build it efficiently. ## 🚀 Key Features * **Unified Agent API**: Simple interface (`Chatbot`) to create agents with memory, tools, and RAG capabilities. * **Tool Management**: Easy-to-use decorators (`@tool`) and class-based toolkits (`AbstractToolkit`) to give your agents capabilities. * **Orchestration & Workflow**: `AgentCrew` for managing multi-agent workflows (Sequential, Parallel, Flow, Loop). * **Advanced Connectivity**: * **A2A (Agent-to-Agent)**: Native protocol for agents to discover and talk to each other. * **MCP (Model Context Protocol)**: Expose your agents as MCP servers or consume external MCP servers. * **OpenAPI Integration**: Consume any OpenAPI specification as a dynamic toolkit (`OpenAPIToolkit`). * **Scheduling**: Built-in task scheduling for agents using the `@schedule` decorator. * **Multi-Provider Support**: Switch seamlessy between OpenAI, Anthropic, Google Gemini, Groq, and more. * **Integrations**: Native support for exposing bots via Telegram, MS Teams, and Slack. --- ## 📦 Installation ```bash pip install ai-parrot ``` For specific provider support (e.g., Anthropic, Google): ```bash pip install "ai-parrot[anthropic,google]" ``` --- ## ⚡ Quick Start Create a simple weather chatbot in just a few lines of code: ```python import asyncio from parrot.bots import Chatbot from parrot.tools import tool # 1. Define a tool @tool def get_weather(location: str) -> str: """Get the current weather for a location.""" return f"The weather in {location} is Sunny, 25°C" async def main(): # 2. Create the Agent bot = Chatbot( name="WeatherBot", llm="openai:gpt-4o", # Provider:Model tools=[get_weather], system_prompt="You are a helpful weather assistant." ) # 3. Configure (loads tools, connects to memory) await bot.configure() # 4. Chat! response = await bot.ask("What's the weather like in Madrid?") print(response) if __name__ == "__main__": asyncio.run(main()) ``` --- ## 🏗️ Architecture AI-Parrot is designed with a modular architecture enabling agents to be both consumers and providers of tools and services. ```mermaid graph TD User["User / Client"] --> API["AgentTalk Handlers"] API --> Bot["Chatbot / BaseBot"] subgraph "Agent Core" Bot --> Memory["Memory / Vector Store"] Bot --> LLM["LLM Client (OpenAI/Anthropic/Etc)"] Bot --> TM["Tool Manager"] end subgraph "Tools & Capabilities" TM --> LocalTools["Local Tools (@tool)"] TM --> Toolkits["Toolkits (OpenAPI/Custom)"] TM --> MCPServer["External MCP Servers"] end subgraph "Connectivity" Bot -.-> A2A["A2A Protocol (Client/Server)"] Bot -.-> MCP["MCP Protocol (Server)"] Bot -.-> Integrations["Telegram / MS Teams"] end subgraph "Orchestration" Crew["AgentCrew"] --> Bot Crew --> OtherBots["Other Agents"] end ``` --- ## 🧩 Core Concepts ### Agents (`Chatbot`) The `Chatbot` class is your main entry point. It handles conversation history, RAG (Retrieval-Augmented Generation), and tool execution loop. ```python bot = Chatbot( name="MyAgent", model="anthropic:claude-3-5-sonnet-20240620", enable_memory=True ) ``` ### Tools #### Functional Tools (`@tool`) The simplest way to create a tool. The docstring and type hints are automatically used to generate the schema for the LLM. ```python from parrot.tools import tool @tool def calculate_vat(amount: float, rate: float = 0.20) -> float: """Calculate VAT for a given amount.""" return amount * rate ``` #### Class-Based Toolkits (`AbstractToolkit`) Group related tools into a reusable class. All public async methods become tools. ```python from parrot.tools import AbstractToolkit class MathToolkit(AbstractToolkit): async def add(self, a: int, b: int) -> int: """Add two numbers.""" return a + b async def multiply(self, a: int, b: int) -> int: """Multiply two numbers.""" return a * b ``` #### OpenAPI Toolkit (`OpenAPIToolkit`) Dynamically generate tools from any OpenAPI/Swagger specification. ```python from parrot.tools.openapi_toolkit import OpenAPIToolkit petstore = OpenAPIToolkit( spec="https://petstore.swagger.io/v2/swagger.json", service="petstore" ) # Now your agent can call petstore_get_pet_by_id, etc. bot = Chatbot(name="PetBot", tools=petstore.get_tools()) ``` ### Orchestration (`AgentCrew`) orchestrate multiple agents to solve complex tasks using `AgentCrew`. **Supported Modes:** * **Sequential**: Agents run one after another, passing context. * **Parallel**: Independent tasks run concurrently. * **Flow**: DAG-based execution defined by dependencies. * **Loop**: Iterative execution until a condition is met. ```python from parrot.bots.orchestration import AgentCrew crew = AgentCrew( name="ResearchTeam", agents=[researcher_agent, writer_agent] ) # Define a Flow # Writer waits for Researcher to finish crew.task_flow(researcher_agent, writer_agent) await crew.run_flow("Research the latest advancements in Quantum Computing") ``` ### Scheduling (`@schedule`) Give your agents agency to run tasks in the background. ```python from parrot.scheduler import schedule, ScheduleType class DailyBot(Chatbot): @schedule(schedule_type=ScheduleType.DAILY, hour=9, minute=0) async def morning_briefing(self): news = await self.ask("Summarize today's top tech news") await self.send_notification(news) ``` --- ## 🔌 Connectivity & Exposure ### Agent-to-Agent (A2A) Protocol Agents can discover and talk to each other using the A2A protocol. **Expose an Agent:** ```python # In your server setup (aiohttp) from parrot.a2a import A2AServer a2a = A2AServer(my_agent) a2a.setup(app, url="https://my-agent.com") ``` **Consume an Agent:** ```python from parrot.a2a import A2AClient async with A2AClient("https://remote-agent.com") as client: response = await client.send_message("Hello from another agent!") ``` ### Model Context Protocol (MCP) **AI-Parrot** has first-class support for MCP. **Consume MCP Servers:** Give your agent access to filesystem, git, or any other MCP server. ```python # In Chatbot config mcp_servers = [ MCPServerConfig( name="filesystem", command="npx", args=["-y", "@modelcontextprotocol/server-filesystem", "/home/user"] ) ] await bot.setup_mcp_servers(mcp_servers) ``` **Expose Agent as MCP Server:** Allow Claude Desktop or other MCP clients to use your agent as a tool. ```python # (Configuration details in documentation) ``` ### Platform Integrations Expose your bots natively to chat platforms defined in your `parrot.conf`: * **Telegram** * **Microsoft Teams** * **Slack** * **WhatsApp** --- ## 🤖 Supported LLM Clients AI-Parrot supports a wide range of LLM providers via `parrot.clients`: * **OpenAI** (`openai`) * **Anthropic** (`anthropic`, `claude`) * **Google Gemini** (`google`) * **Groq** (`groq`) * **X.AI** (`grok`) * **HuggingFace** (`hf`) * **Ollama/Local** (via OpenAI compatible endpoint) --- ## 🤝 Community & Support * **Issues**: [GitHub Tracker](https://github.com/phenobarbital/ai-parrot/issues) * **Discussion**: [GitHub Discussions](https://github.com/phenobarbital/ai-parrot/discussions) * **Contribution**: Pull requests are welcome! Please read `CONTRIBUTING.md`. --- *Built with ❤️ by the Navigator Team*
text/markdown
null
Jesus Lara <jesuslara@phenobarbital.info>
null
null
null
asyncio, asyncpg, aioredis, aiomcache, artificial intelligence, ai, chatbot, agents
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "Operating System :: POSIX :: Linux", "Environment :: Web Environment", "Topic :: Software Development :: Build Tools", "Topic :: Software Development :: Libraries :: Python Modules", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3 :: Only", "Framework :: AsyncIO", "Typing :: Typed" ]
[]
null
null
>=3.10.1
[]
[]
[]
[ "Cython==3.0.11", "faiss-cpu>=1.9.0", "jq==1.7.0", "rank_bm25==0.2.2", "tabulate==0.9.0", "sentencepiece==0.2.1", "markdown2==2.5.4", "psycopg-binary==3.2.6", "python-datamodel>=0.10.17", "backoff==2.2.1", "asyncdb>=2.11.6", "google-cloud-bigquery>=3.30.0", "numexpr==2.10.2", "fpdf==1.7.2", "python-docx==1.1.2", "typing-extensions<5,>=4.14.1", "navconfig[default]>=1.7.13", "navigator-auth>=0.17.2", "navigator-session>=0.6.5", "navigator-api[locale,uvloop]>=2.13.5", "matplotlib==3.10.0", "seaborn==0.13.2", "pydub==0.25.1", "click>=8.1.7", "async-notify[all]>=1.4.2", "markitdown>=0.1.2", "ddgs>=9.5.2", "xmltodict>=0.14.2", "pytesseract>=0.3.13", "python-statemachine==2.5.0", "aiohttp-swagger3==0.10.0", "PyYAML>=6.0.2", "python-arango-async==1.2.0", "asyncdb[arangodb,bigquery,boto3,influxdb,mongodb]>=2.12.0", "brotli==1.2.0", "urllib3==2.6.3", "aioquic==1.3.0", "pylsqpack==0.3.23", "aiohttp-sse-client==0.2.1", "prance>=25.4.8.0", "openapi-schema-validator==0.6.3", "praw>=7.8.1", "weasyprint==68.0", "apscheduler==3.11.2", "pydantic==2.12.5", "aiohttp-cors>=0.8.1", "querysource>=3.17.10", "pywa>=3.8.0", "sentence_transformers==5.0.0; extra == \"agents\"", "yfinance==0.2.54; extra == \"agents\"", "youtube_search==2.1.2; extra == \"agents\"", "wikipedia==1.4.0; extra == \"agents\"", "mediawikiapi==1.2; extra == \"agents\"", "pyowm==3.3.0; extra == \"agents\"", "stackapi==0.3.1; extra == \"agents\"", "duckduckgo-search==8.1.1; extra == \"agents\"", "google-search-results==2.4.2; extra == \"agents\"", "google-api-python-client>=2.151.0; extra == \"agents\"", "networkx>=3.0; extra == \"agents\"", "decorator>=5; extra == \"agents\"", "autoviz==0.1.905; extra == \"agents\"", "spacy==3.8.6; extra == \"agents\"", "html2text==2025.4.15; extra == \"agents\"", "httpx-sse==0.4.1; extra == \"agents\"", "mcp==1.15.0; extra == \"agents\"", "sse-starlette==3.0.2; extra == \"agents\"", "requests-oauthlib==2.0.0; extra == \"agents\"", "undetected-chromedriver==3.5.5; extra == \"agents\"", "selenium==4.35.0; extra == \"agents\"", "playwright==1.52.0; extra == \"agents\"", "streamlit==1.50.0; extra == \"agents\"", "jira==3.10.5; extra == \"agents\"", "arxiv==2.2.0; extra == \"agents\"", "docker==7.1.0; extra == \"agents\"", "aiogoogle==5.17.0; extra == \"agents\"", "rq==2.6.0; extra == \"agents\"", "zeep[async]==4.3.1; extra == \"agents\"", "branca==0.8.2; extra == \"agents\"", "folium==0.20.0; extra == \"agents\"", "webdriver-manager==4.0.2; extra == \"agents\"", "prophet==1.2.1; extra == \"agents\"", "folium==0.20.0; extra == \"agents\"", "opensearch-py==3.1.0; extra == \"agents\"", "cairosvg>=2.7; extra == \"agents\"", "python-pptx==1.0.2; extra == \"agents\"", "markdownify==1.1.0; extra == \"agents\"", "python-docx==1.1.2; extra == \"agents\"", "pymupdf==1.26.3; extra == \"agents\"", "pymupdf4llm==0.0.27; extra == \"agents\"", "pdf4llm==0.0.27; extra == \"agents\"", "alpaca-py>=0.43.2; extra == \"agents\"", "defillama-sdk>=0.1.0; extra == \"agents\"", "pandas-ta-classic>=0.3.59; extra == \"agents\"", "TA-Lib>=0.4.32; extra == \"agents\"", "aioimaplib>=1.1.0; extra == \"agents\"", "gmqtt>=0.6.15; extra == \"agents\"", "azure-identity>=1.18.0; extra == \"agents\"", "msgraph-sdk>=1.8.0; extra == \"agents\"", "microsoft-kiota-authentication-azure>=1.2.0; extra == \"agents\"", "jinja2>=3.1; extra == \"agents\"", "xhtml2pdf>=0.2.17; extra == \"agents\"", "matplotlib>=3.7; extra == \"charts\"", "cairosvg>=2.7; extra == \"charts\"", "svglib>=1.5; extra == \"charts\"", "reportlab>=4.0; extra == \"charts\"", "yfinance==0.2.54; extra == \"agents-lite\"", "youtube_search==2.1.2; extra == \"agents-lite\"", "wikipedia==1.4.0; extra == \"agents-lite\"", "mediawikiapi==1.2; extra == \"agents-lite\"", "pyowm==3.3.0; extra == \"agents-lite\"", "stackapi==0.3.1; extra == \"agents-lite\"", "duckduckgo-search==8.1.1; extra == \"agents-lite\"", "google-search-results==2.4.2; extra == \"agents-lite\"", "google-api-python-client>=2.151.0; extra == \"agents-lite\"", "networkx>=3.0; extra == \"agents-lite\"", "decorator>=5; extra == \"agents-lite\"", "html2text==2025.4.15; extra == \"agents-lite\"", "httpx-sse==0.4.1; extra == \"agents-lite\"", "mcp==1.15.0; extra == \"agents-lite\"", "sse-starlette==3.0.2; extra == \"agents-lite\"", "requests-oauthlib==2.0.0; extra == \"agents-lite\"", "jira==3.10.5; extra == \"agents-lite\"", "arxiv==2.2.0; extra == \"agents-lite\"", "docker==7.1.0; extra == \"agents-lite\"", "aiogoogle==5.17.0; extra == \"agents-lite\"", "rq==2.6.0; extra == \"agents-lite\"", "zeep[async]==4.3.1; extra == \"agents-lite\"", "branca==0.8.2; extra == \"agents-lite\"", "folium==0.20.0; extra == \"agents-lite\"", "opensearch-py==3.1.0; extra == \"agents-lite\"", "mammoth==1.8.0; extra == \"loaders\"", "pytube==15.0.0; extra == \"loaders\"", "youtube_transcript_api==1.0.3; extra == \"loaders\"", "yt-dlp==2025.8.22; extra == \"loaders\"", "ebooklib>=0.19; extra == \"loaders\"", "whisperx==3.4.2; extra == \"loaders\"", "av==15.1.0; extra == \"loaders\"", "resemblyzer==0.1.4; extra == \"loaders\"", "pyannote-audio==3.4.0; extra == \"loaders\"", "pyannote-core==5.0.0; extra == \"loaders\"", "pyannote-database==5.1.3; extra == \"loaders\"", "pyannote-metrics==3.2.1; extra == \"loaders\"", "pyannote-pipeline==3.0.1; extra == \"loaders\"", "pytorch-lightning==2.5.5; extra == \"loaders\"", "pytorch-metric-learning==2.9.0; extra == \"loaders\"", "nvidia-cudnn-cu12==9.1.0.70; extra == \"loaders\"", "moviepy==2.2.1; extra == \"loaders\"", "decorator>=5; extra == \"loaders\"", "ffmpeg==1.4; extra == \"loaders\"", "paddleocr==3.2.0; extra == \"loaders\"", "sentence-transformers>=5.0.0; extra == \"embeddings\"", "tiktoken==0.9.0; extra == \"embeddings\"", "chromadb==0.6.3; extra == \"embeddings\"", "bm25s[full]==0.2.14; extra == \"embeddings\"", "simsimd>=4.3.1; extra == \"embeddings\"", "tokenizers<=0.21.1,>=0.20.0; extra == \"embeddings\"", "safetensors>=0.4.3; extra == \"embeddings\"", "torch==2.6.0; extra == \"ml-heavy\"", "torchaudio==2.6.0; extra == \"ml-heavy\"", "numpy<2.2,>=2.1; extra == \"ml-heavy\"", "accelerate==0.34.2; extra == \"ml-heavy\"", "bitsandbytes==0.44.1; extra == \"ml-heavy\"", "datasets>=3.0.2; extra == \"ml-heavy\"", "transformers<=4.51.3,>=4.51.1; extra == \"ml-heavy\"", "tensorflow>=2.19.1; extra == \"ml-heavy\"", "tf-keras==2.19.0; extra == \"ml-heavy\"", "opencv-python==4.10.0.84; extra == \"ml-heavy\"", "ai-parrot[embeddings,ml-heavy]; extra == \"vectors\"", "google-genai>=1.61.0; extra == \"mcp\"", "openai==2.8.1; extra == \"mcp\"", "yfinance==0.2.54; extra == \"mcp\"", "youtube_search==2.1.2; extra == \"mcp\"", "wikipedia==1.4.0; extra == \"mcp\"", "mediawikiapi==1.2; extra == \"mcp\"", "pyowm==3.3.0; extra == \"mcp\"", "stackapi==0.3.1; extra == \"mcp\"", "duckduckgo-search==8.1.1; extra == \"mcp\"", "google-search-results==2.4.2; extra == \"mcp\"", "google-api-python-client>=2.151.0; extra == \"mcp\"", "networkx>=3.0; extra == \"mcp\"", "decorator>=5; extra == \"mcp\"", "html2text==2025.4.15; extra == \"mcp\"", "httpx-sse==0.4.1; extra == \"mcp\"", "mcp==1.15.0; extra == \"mcp\"", "sse-starlette==3.0.2; extra == \"mcp\"", "requests-oauthlib==2.0.0; extra == \"mcp\"", "jira==3.10.5; extra == \"mcp\"", "arxiv==2.2.0; extra == \"mcp\"", "docker==7.1.0; extra == \"mcp\"", "aiogoogle==5.17.0; extra == \"mcp\"", "rq==2.6.0; extra == \"mcp\"", "zeep[async]==4.3.1; extra == \"mcp\"", "branca==0.8.2; extra == \"mcp\"", "folium==0.20.0; extra == \"mcp\"", "opensearch-py==3.1.0; extra == \"mcp\"", "torchvision==0.21.0; extra == \"images\"", "timm==1.0.15; extra == \"images\"", "ultralytics==8.3.179; extra == \"images\"", "albumentations==2.0.6; extra == \"images\"", "filetype==1.2.0; extra == \"images\"", "imagehash==4.3.1; extra == \"images\"", "pgvector==0.4.1; extra == \"images\"", "pyheif==0.8.0; extra == \"images\"", "exif==1.6.1; extra == \"images\"", "pillow-avif-plugin==1.5.2; extra == \"images\"", "pillow-heif==0.22.0; extra == \"images\"", "python-xmp-toolkit==2.0.2; extra == \"images\"", "exifread==3.5.1; extra == \"images\"", "transformers<=4.51.3,>=4.51.1; extra == \"images\"", "ffmpeg==1.4; extra == \"images\"", "holoviews==1.21.0; extra == \"images\"", "bokeh==3.7.3; extra == \"images\"", "pandas-bokeh==0.5.5; extra == \"images\"", "plotly==5.22.0; extra == \"images\"", "ipywidgets==8.1.0; extra == \"images\"", "altair==5.5.0; extra == \"images\"", "whisperx==3.4.2; extra == \"whisperx\"", "av==15.1.0; extra == \"whisperx\"", "torch==2.6.0; extra == \"whisperx\"", "torchaudio==2.6.0; extra == \"whisperx\"", "torchvision==0.21.0; extra == \"whisperx\"", "pyannote-audio==3.4.0; extra == \"whisperx\"", "pyannote-core==5.0.0; extra == \"whisperx\"", "pyannote-database==5.1.3; extra == \"whisperx\"", "pyannote-metrics==3.2.1; extra == \"whisperx\"", "pyannote-pipeline==3.0.1; extra == \"whisperx\"", "pytorch-lightning==2.5.5; extra == \"whisperx\"", "pytorch-metric-learning==2.9.0; extra == \"whisperx\"", "nvidia-cudnn-cu12==9.1.0.70; extra == \"whisperx\"", "torch-audiomentations==0.12.0; extra == \"whisperx\"", "torch-pitch-shift==1.2.5; extra == \"whisperx\"", "torchmetrics==1.8.2; extra == \"whisperx\"", "anthropic[aiohttp]==0.61.0; extra == \"anthropic\"", "claude-agent-sdk>=0.1.0; extra == \"anthropic\"", "openai==2.8.1; extra == \"openai\"", "tiktoken==0.9.0; extra == \"openai\"", "google-api-python-client<=2.177.0,>=2.166.0; extra == \"google\"", "google-cloud-texttospeech==2.27.0; extra == \"google\"", "google-genai>=1.61.0; extra == \"google\"", "google-cloud-aiplatform==1.110.0; extra == \"google\"", "groq==0.33.0; extra == \"groq\"", "google-genai>=1.61.0; extra == \"llms\"", "openai==2.8.1; extra == \"llms\"", "groq==0.33.0; extra == \"llms\"", "anthropic[aiohttp]==0.61.0; extra == \"llms\"", "claude-agent-sdk>=0.1.0; extra == \"llms\"", "xai-sdk>=0.1.0; extra == \"llms\"", "querysource>=3.17.9; extra == \"integrations\"", "async-notify[all]>=1.5.2; extra == \"integrations\"", "azure-teambots>=0.1.1; extra == \"integrations\"", "pymilvus==2.4.8; extra == \"milvus\"", "milvus==2.3.5; extra == \"milvus\"", "chroma==0.2.0; extra == \"chroma\"", "ydata-profiling==4.16.1; extra == \"eda\"", "sweetviz==2.1.4; extra == \"eda\"", "pytector[gguf]==0.2.0; extra == \"security\"", "xai-sdk>=0.1.0; extra == \"xai\"", "gunicorn>=23.0.0; extra == \"deploy\"", "docling[tesserocr]>=2.31.1; extra == \"docling\"", "mautrix>=0.20; extra == \"matrix\"", "python-olm>=3.2.16; extra == \"matrix\"", "ai-parrot[agents,images,integrations,llms,loaders,vectors]; extra == \"all\"", "ai-parrot[agents-lite,embeddings,integrations,llms]; extra == \"all-fast\"", "pytest>=7.2.2; extra == \"dev\"", "pytest-asyncio==1.2.0; extra == \"dev\"", "pytest-xdist==3.3.1; extra == \"dev\"", "pytest-assume==2.4.3; extra == \"dev\"", "pytest-mock==3.15.1; extra == \"dev\"", "black; extra == \"dev\"", "pylint; extra == \"dev\"", "mypy; extra == \"dev\"", "coverage; extra == \"dev\"", "maturin==1.9.6; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://github.com/phenobarbital/ai-parrot", "Source, https://github.com/phenobarbital/ai-parrot", "Tracker, https://github.com/phenobarbital/ai-parrot/issues", "Documentation, https://github.com/phenobarbital/ai-parrot/", "Funding, https://paypal.me/phenobarbital", "Say Thanks!, https://saythanks.io/to/phenobarbital" ]
twine/6.2.0 CPython/3.12.3
2026-02-21T00:39:20.948078
ai_parrot-0.23.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl
3,568,512
e5/d6/6efb8ad8934047b65c2c7d9b3e2a8dd158b5e2daeec1ccc257e3dfa11c9c/ai_parrot-0.23.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl
cp312
bdist_wheel
null
false
d2ac008387d7ce523bde4b3536f14f4b
02c1d27a6d1517956be2d00ecc8f3b7766995672aa2fe4c6d0927092d2ae8b5e
e5d66efb8ad8934047b65c2c7d9b3e2a8dd158b5e2daeec1ccc257e3dfa11c9c
MIT
[ "LICENSE" ]
418
2.4
Geode-Conversion
6.5.14
Conversion module for Geode-solutions OpenGeode modules
<h1 align="center">Geode-Conversion<sup><i>by Geode-solutions</i></sup></h1> <h3 align="center">Conversion OpenGeode module</h3> <p align="center"> <img src="https://github.com/Geode-solutions/Geode-Conversion_private/workflows/CI/badge.svg" alt="Build Status"> <img src="https://github.com/Geode-solutions/Geode-Conversion_private/workflows/CD/badge.svg" alt="Deploy Status"> <img src="https://codecov.io/gh/Geode-solutions/Geode-Conversion_private/branch/master/graph/badge.svg" alt="Coverage Status"> <img src="https://img.shields.io/github/release/Geode-solutions/Geode-Conversion_private.svg" alt="Version"> </p> <p align="center"> <img src="https://img.shields.io/static/v1?label=Windows&logo=windows&logoColor=white&message=support&color=success" alt="Windows support"> <img src="https://img.shields.io/static/v1?label=Ubuntu&logo=Ubuntu&logoColor=white&message=support&color=success" alt="Ubuntu support"> <img src="https://img.shields.io/static/v1?label=Red%20Hat&logo=Red-Hat&logoColor=white&message=support&color=success" alt="Red Hat support"> </p> <p align="center"> <img src="https://img.shields.io/badge/C%2B%2B-11-blue.svg" alt="Language"> <img src="https://img.shields.io/badge/license-MIT-blue.svg" alt="License"> <img src="https://img.shields.io/badge/%20%20%F0%9F%93%A6%F0%9F%9A%80-semantic--release-e10079.svg" alt="Semantic-release"> <a href="https://geode-solutions.com/#slack"> <img src="https://opengeode-slack-invite.herokuapp.com/badge.svg" alt="Slack invite"> </a> Copyright (c) 2019 - 2026, Geode-solutions
text/markdown
null
Geode-solutions <contact@geode-solutions.com>
null
null
Proprietary
null
[]
[]
null
null
<3.13,>=3.9
[]
[]
[]
[ "geode-common==33.*,>=33.19.0", "opengeode-core==15.*,>=15.31.5" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.9.13
2026-02-21T00:39:03.085531
geode_conversion-6.5.14-cp39-cp39-win_amd64.whl
2,058,988
79/75/61d7837cd4b6998f99ab4950640881ff0f179335273e0ab67f18b7120072/geode_conversion-6.5.14-cp39-cp39-win_amd64.whl
cp39
bdist_wheel
null
false
07349c6231c7fd903be40db9f687e25d
f0a087ece1662965a14228fa132ab7cdb604368a374cfd91b3e13dee212e7676
797561d7837cd4b6998f99ab4950640881ff0f179335273e0ab67f18b7120072
null
[]
0
2.1
kuhl-haus-mdp
0.2.24
Market data processing pipeline for stock market scanner
.. image:: https://img.shields.io/github/license/kuhl-haus/kuhl-haus-mdp :alt: License :target: https://github.com/kuhl-haus/kuhl-haus-mdp/blob/mainline/LICENSE.txt .. image:: https://img.shields.io/pypi/v/kuhl-haus-mdp.svg :alt: PyPI :target: https://pypi.org/project/kuhl-haus-mdp/ .. image:: https://static.pepy.tech/badge/kuhl-haus-mdp/month :alt: Downloads :target: https://pepy.tech/project/kuhl-haus-mdp .. image:: https://github.com/kuhl-haus/kuhl-haus-mdp/actions/workflows/publish-to-pypi.yml/badge.svg :alt: Build Status :target: https://github.com/kuhl-haus/kuhl-haus-mdp/actions/workflows/publish-to-pypi.yml .. image:: https://github.com/kuhl-haus/kuhl-haus-mdp/actions/workflows/codeql.yml/badge.svg :alt: CodeQL Advanced :target: https://github.com/kuhl-haus/kuhl-haus-mdp/actions/workflows/codeql.yml .. image:: https://codecov.io/gh/kuhl-haus/kuhl-haus-mdp/branch/mainline/graph/badge.svg :alt: codecov :target: https://codecov.io/gh/kuhl-haus/kuhl-haus-mdp .. image:: https://img.shields.io/github/issues/kuhl-haus/kuhl-haus-mdp :alt: GitHub issues :target: https://github.com/kuhl-haus/kuhl-haus-mdp/issues .. image:: https://img.shields.io/github/issues-pr/kuhl-haus/kuhl-haus-mdp :alt: GitHub pull requests :target: https://github.com/kuhl-haus/kuhl-haus-mdp/pulls .. image:: https://readthedocs.org/projects/kuhl-haus-mdp/badge/?version=latest :alt: Documentation :target: https://kuhl-haus-mdp.readthedocs.io/en/latest/ | ============== kuhl-haus-mdp ============== Market data processing library. Overview ======== The Kuhl Haus Market Data Platform (MDP) is a distributed system for collecting, processing, and serving real-time market data. Built on Kubernetes and leveraging microservices architecture, MDP provides scalable infrastructure for financial data analysis and visualization. Key Features ------------ - Real-time market data ingestion and processing - Scalable microservices architecture - Automated deployment with Ansible and Kubernetes - Multi-environment support (development, staging, production) - OAuth integration for secure authentication - Redis-based caching layer for performance Code Organization ----------------- The platform consists of four main packages: - **Market data processing library** (`kuhl-haus-mdp <https://github.com/kuhl-haus/kuhl-haus-mdp>`_) - Core library with shared data processing logic - **Backend Services** (`kuhl-haus-mdp-servers <https://github.com/kuhl-haus/kuhl-haus-mdp-servers>`_) - Market data listener, processor, and widget service - **Frontend Application** (`kuhl-haus-mdp-app <https://github.com/kuhl-haus/kuhl-haus-mdp-app>`_) - Web-based user interface and API - **Deployment Automation** (`kuhl-haus-mdp-deployment <https://github.com/kuhl-haus/kuhl-haus-mdp-deployment>`_) - Docker Compose, Ansible playbooks and Kubernetes manifests for environment provisioning Additional Resources -------------------- 📖 **Blog Series:** - `Part 1: Why I Built It <https://the.oldschool.engineer/what-i-built-after-quitting-amazon-spoiler-its-a-stock-scanner-28fc3b6d9be0>`_ - `Part 2: How to Run It <https://the.oldschool.engineer/what-i-built-after-quitting-amazon-spoiler-its-a-stock-scanner-part-2-94e445914951>`_ - `Part 3: How to Deploy It <https://the.oldschool.engineer/what-i-built-after-quitting-amazon-spoiler-its-a-stock-scanner-part-3-eab7d9bbf5f7>`_ - `Part 4: Evolution from Prototype to Production <https://the.oldschool.engineer/what-i-built-after-quitting-amazon-spoiler-its-a-stock-scanner-part-4-408779a1f3f2>`_ Components Summary ================== .. figure:: Market_Data_Processing_C4.png :align: center :alt: Market Data Platform Context Diagram Market Data Platform Context Diagram Data Plane Components ---------------------- **Market Data Listener (MDL)** WebSocket client connecting to Massive.com, routing events to appropriate queues with minimal processing overhead. **Market Data Queues (MDQ)** RabbitMQ-based FIFO queues with 5-second TTL, buffering high-velocity streams for distributed processing. **Market Data Processor (MDP)** Horizontally-scalable event processors with semaphore-based concurrency (500 concurrent tasks), delegating to pluggable analyzers. **Market Data Cache (MDC)** Redis-backed cache layer with TTL policies (5s-24h), atomic operations, and pub/sub distribution. **Widget Data Service (WDS)** WebSocket-to-Redis bridge providing real-time streaming to client applications with fan-out pattern. Control Plane ------------- **Service Control Plane (SCP)** OAuth authentication, SPA serving, runtime controls, and management API (external repository: kuhl-haus-mdp-app). Observability ------------- All components emit OpenTelemetry traces/metrics and structured JSON logs for Kubernetes/OpenObserve integration. Deployment Model ================ The platform deploys to Kubernetes with independent scaling per component: - **Data plane**: Internal network only (MDL, MDQ, MDP, MDC) - **Client interface**: Exposed to client networks (WDS) - **Control plane**: External access (SCP) All components run as Docker containers with automated deployment via Ansible playbooks and Kubernetes manifests (kuhl-haus-mdp-deployment repository). Component Descriptions ====================== .. figure:: architecture.svg :align: center :alt: Market Data Platform Component Architecture Market Data Platform Component Architecture Market Data Listener (MDL) --------------------------- The MDL performs minimal processing on the messages. MDL inspects the message type for selecting the appropriate serialization method and destination queue. MDL implementations may vary as new MDS become available (for example, news). MDL runs as a container and scales independently of other components. The MDL should not be accessible outside the data plane local network. Code Libraries ~~~~~~~~~~~~~~ - **MassiveDataListener** (``components/massive_data_listener.py``) - WebSocket client wrapper for Massive.com with persistent connection management and market-aware reconnection logic - **MassiveDataQueues** (``components/massive_data_queues.py``) - Multi-channel RabbitMQ publisher routing messages by event type with concurrent batch publishing (100 msg/frame) - **WebSocketMessageSerde** (``helpers/web_socket_message_serde.py``) - Serialization/deserialization for Massive WebSocket messages to/from JSON - **QueueNameResolver** (``helpers/queue_name_resolver.py``) - Event type to queue name routing logic Market Data Queues (MDQ) ------------------------- **Purpose:** Buffer high-velocity market data stream for server-side processing with aggressive freshness controls - **Queue Type:** FIFO with TTL (5-second max message age) - **Cleanup Strategy:** Discarded when TTL expires - **Message Format:** Timestamped JSON preserving original Massive.com structure - **Durability:** Non-persistent messages (speed over reliability for real-time data) - **Independence:** Queues operate completely independently - one queue per subscription - **Technology:** RabbitMQ The MDQ should not be accessible outside the data plane local network. Code Libraries ~~~~~~~~~~~~~~ - **MassiveDataQueues** (``components/massive_data_queues.py``) - Queue setup, per-queue channel management, and message publishing with NOT_PERSISTENT delivery mode - **MassiveDataQueue** enum (``enum/massive_data_queue.py``) - Queue name constants for routing (AGGREGATE, TRADES, QUOTES, HALTS, UNKNOWN) Market Data Processors (MDP) ----------------------------- The purpose of the MDP is to process raw real-time market data and delegate processing to data-specific handlers. This separation of concerns allows MDPs to handle any type of data and simplifies horizontal scaling. The MDP stores its processed results in the Market Data Cache (MDC). The MDP: - Hydrates the in-memory cache on MDC - Processes market data - Publishes messages to pub/sub channels - Maintains cache entries in MDC MDPs runs as containers and scale independently of other components. The MDPs should not be accessible outside the data plane local network. Code Libraries ~~~~~~~~~~~~~~ - **MassiveDataProcessor** (``components/massive_data_processor.py``) - RabbitMQ consumer with semaphore-based concurrency control for high-throughput scenarios (1,000+ events/sec) - **MarketDataScanner** (``components/market_data_scanner.py``) - Redis pub/sub consumer with pluggable analyzer pattern for sequential message processing - **Analyzers** (``analyzers/``) - **MassiveDataAnalyzer** (``massive_data_analyzer.py``) - Stateless event router dispatching by event type - **LeaderboardAnalyzer** (``leaderboard_analyzer.py``) - Redis sorted set leaderboards (volume, gappers, gainers) with day/market boundary resets and distributed throttling - **TopTradesAnalyzer** (``top_trades_analyzer.py``) - Redis List-based trade history with sliding window (last 1,000 trades/symbol) and aggregated statistics - **TopStocksAnalyzer** (``top_stocks.py``) - In-memory leaderboard prototype (legacy, single-instance) - **MarketDataAnalyzerResult** (``data/market_data_analyzer_result.py``) - Result envelope for analyzer output with cache/publish metadata - **ProcessManager** (``helpers/process_manager.py``) - Multiprocess orchestration for async workers with OpenTelemetry context propagation Market Data Cache (MDC) ------------------------ **Purpose:** In-memory data store for serialized processed market data. - **Cache Type:** In-memory persistent or with TTL - **Queue Type:** pub/sub - **Technology:** Redis The MDC should not be accessible outside the data plane local network. Code Libraries ~~~~~~~~~~~~~~ - **MarketDataCache** (``components/market_data_cache.py``) - Redis cache-aside layer for Massive.com API with TTL policies, negative caching, and specialized metric methods (snapshot, avg volume, free float) - **MarketDataCacheKeys** enum (``enum/market_data_cache_keys.py``) - Internal Redis cache key patterns and templates - **MarketDataCacheTTL** enum (``enum/market_data_cache_ttl.py``) - TTL values balancing freshness vs. API quotas vs. memory pressure (5s for trades, 24h for reference data) - **MarketDataPubSubKeys** enum (``enum/market_data_pubsub_keys.py``) - Redis pub/sub channel names for external consumption Widget Data Service (WDS) -------------------------- **Purpose:** 1. WebSocket interface provides access to processed market data for client-side code 2. Is the network-layer boundary between clients and the data that is available on the data plane WDS runs as a container and scales independently of other components. WDS is the only data plane component that should be exposed to client networks. Code Libraries ~~~~~~~~~~~~~~ - **WidgetDataService** (``components/widget_data_service.py``) - WebSocket-to-Redis bridge with fan-out pattern, lazy task initialization, wildcard subscription support, and lock-protected subscription management - **MarketDataCache** (``components/market_data_cache.py``) - Snapshot retrieval for initial state before streaming Service Control Plane (SCP) ---------------------------- **Purpose:** 1. Authentication and authorization 2. Serve static and dynamic content via py4web 3. Serve SPA to authenticated clients 4. Injects authentication token and WDS url into SPA environment for authenticated access to WDS 5. Control plane for managing application components at runtime 6. API for programmatic access to service controls and instrumentation. The SCP requires access to the data plane network for API access to data plane components. The SCP code is in the `kuhl-haus/kuhl-haus-mdp-app <https://github.com/kuhl-haus/kuhl-haus-mdp-app>`_ repo. Miscellaneous Code Libraries ----------------------------- - **Observability** (``helpers/observability.py``) - OpenTelemetry tracer/meter factory for distributed tracing and metrics - **StructuredLogging** (``helpers/structured_logging.py``) - JSON logging for K8s/OpenObserve with dev mode support - **Utils** (``helpers/utils.py``) - API key resolution (MASSIVE_API_KEY → POLYGON_API_KEY → file) and TickerSnapshot serialization
text/x-rst
null
Tom Pounders <git@oldschool.engineer>
null
null
The MIT License (MIT) Copyright (c) 2025 Tom Pounders Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
null
[ "Development Status :: 4 - Beta", "Programming Language :: Python" ]
[]
null
null
>=3.14
[]
[]
[]
[ "aiohttp", "aio-pika", "fastapi", "massive", "opentelemetry-api", "opentelemetry-sdk", "opentelemetry-exporter-otlp", "pydantic-settings", "python-dotenv", "python-json-logger>=2.0.0", "redis", "tenacity", "uvicorn[standard]", "websockets", "setuptools; extra == \"testing\"", "pdm-backend; extra == \"testing\"", "pytest; extra == \"testing\"", "pytest-cov; extra == \"testing\"", "pytest-asyncio; extra == \"testing\"" ]
[]
[]
[]
[ "Homepage, https://github.com/kuhl-haus/kuhl-haus-mdp", "Documentation, https://kuhl-haus-mdp.readthedocs.io", "Source, https://github.com/kuhl-haus/kuhl-haus-mdp", "Changelog, https://github.com/kuhl-haus/kuhl-haus-mdp/blob/mainline/CHANGELOG.rst", "Tracker, https://github.com/kuhl-haus/kuhl-haus-mdp/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:38:51.360904
kuhl_haus_mdp-0.2.24.tar.gz
81,474
7c/d8/0764ea0672b807b76ed25792f0ada79d391c3f19faba72f4a58a2454daeb/kuhl_haus_mdp-0.2.24.tar.gz
source
sdist
null
false
f1ae12dcc343828ecda584c307d45352
ca09bb9517bed23efa423165d2ab3758ce8c00d72d5c1228e955a9e6b95c6002
7cd80764ea0672b807b76ed25792f0ada79d391c3f19faba72f4a58a2454daeb
null
[]
219
2.1
cdk-docker-image-deployment
0.0.953
This module allows you to copy docker image assets to a repository you control. This can be necessary if you want to build a Docker image in one CDK app and consume it in a different app or outside the CDK.
## CDK Docker Image Deployment This module allows you to copy docker image assets to a repository you control. This can be necessary if you want to build a Docker image in one CDK app and consume it in a different app or outside the CDK, or if you want to apply a lifecycle policy to all images of a part of your application. ### Getting Started Below is a basic example for how to use the `DockerImageDeployment` API: ```python import * as ecr from 'aws-cdk-lib/aws-ecr'; import * as imagedeploy from 'cdk-docker-image-deployment'; const repo = ecr.Repository.fromRepositoryName(this, 'MyRepository', 'myrepository'); new imagedeploy.DockerImageDeployment(this, 'ExampleImageDeploymentWithTag', { source: imagedeploy.Source.directory('path/to/directory'), destination: imagedeploy.Destination.ecr(repo, { tag: 'myspecialtag', }), }); ``` ### Currently Supported Sources * `Source.directory()`: Supply a path to a local docker image as source. > Don't see a source listed? See if there is an open [issue](https://github.com/cdklabs/cdk-docker-image-deployment/issues) > or [PR](https://github.com/cdklabs/cdk-docker-image-deployment/pulls) already. If not, please open an issue asking for it > or better yet, submit a contribution! ### Currently Supported Destinations * `Destination.ecr(repo, options)`: Send your docker image to an ECR repository in your stack's account. > Don't see a destination listed? See if there is an open [issue](https://github.com/cdklabs/cdk-docker-image-deployment/issues) > or [PR](https://github.com/cdklabs/cdk-docker-image-deployment/pulls) already. If not, please open an issue asking for it > or better yet, submit a contribution! ### Under the Hood 1. When this stack is deployed (either via cdk deploy or via CI/CD), the contents of the local Docker image will be archived and uploaded to an intermediary assets ECR Repository using the cdk-assets mechanism. 2. The `DockerImageDeployment` construct synthesizes a CodeBuild Project which uses docker to pull the image from the intermediary repository, tag the image if a tag is provided, and push the image to the destination repository. 3. The deployment will wait until the CodeBuild Project completes successfully before finishing. The architecture of this construct can be seen here: ![Construct-Architecture](https://user-images.githubusercontent.com/36202692/187282269-7ab29d3e-192f-470f-9123-5dbb62d9dac3.jpg) ## Security See [CONTRIBUTING](CONTRIBUTING.md#security-issue-notifications) for more information. ## License This project is licensed under the Apache-2.0 License.
text/markdown
Parker Scanlon
null
null
null
Apache-2.0
null
[ "Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: JavaScript", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Typing :: Typed", "Development Status :: 5 - Production/Stable", "License :: OSI Approved" ]
[]
https://github.com/cdklabs/cdk-docker-image-deployment#readme
null
~=3.9
[]
[]
[]
[ "aws-cdk-lib<3.0.0,>=2.24.0", "constructs<11.0.0,>=10.5.1", "jsii<2.0.0,>=1.126.0", "publication>=0.0.3", "typeguard==2.13.3" ]
[]
[]
[]
[ "Source, https://github.com/cdklabs/cdk-docker-image-deployment.git" ]
twine/6.1.0 CPython/3.14.2
2026-02-21T00:38:13.635983
cdk_docker_image_deployment-0.0.953.tar.gz
13,104,057
47/b6/3cfd31c54e9e8f49683a0c8dbcfd31be1dea8e7aa3dc904cd5ea61337870/cdk_docker_image_deployment-0.0.953.tar.gz
source
sdist
null
false
292027d531400389e0a7602ef7c3a2f1
a18e144943bde06ada3d3cad1f746edb55d24b4657e223fef5765c83a15f778b
47b63cfd31c54e9e8f49683a0c8dbcfd31be1dea8e7aa3dc904cd5ea61337870
null
[]
222
2.4
tokencostauto
0.1.516
To calculate token and translated USD cost of string and message calls to OpenAI, for example when used by AI agents
<p align="center"> <img src="https://raw.githubusercontent.com/AgentOps-AI/tokencost/main/tokencost.png" height="300" alt="Tokencost" /> </p> <p align="center"> <em>Clientside token counting + price estimation for LLM apps and AI agents.</em> </p> <p align="center"> <a href="https://pypi.org/project/tokencostauto/" target="_blank"> <img alt="Python" src="https://img.shields.io/badge/python-3670A0?style=for-the-badge&logo=python&logoColor=ffdd54" /> <img alt="Version" src="https://img.shields.io/pypi/v/tokencostauto?style=for-the-badge&color=3670A0"> </a> </p> <p align="center"> <a href="https://twitter.com/agentopsai/">🐦 Twitter</a> <span>&nbsp;&nbsp;•&nbsp;&nbsp;</span> <a href="https://discord.com/invite/FagdcwwXRR">📢 Discord</a> <span>&nbsp;&nbsp;•&nbsp;&nbsp;</span> <a href="https://agentops.ai/?tokencostauto">🖇️ AgentOps</a> </p> # TokenCost [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/licenses/MIT) ![PyPI - Version](https://img.shields.io/pypi/v/tokencost) [![X (formerly Twitter) Follow](https://img.shields.io/twitter/follow/AgentOpsAI)](https://x.com/agentopsai) Tokencost helps calculate the USD cost of using major Large Language Model (LLMs) APIs by calculating the estimated cost of prompts and completions. Building AI agents? Check out [AgentOps](https://agentops.ai/?tokencostauto) ### Features * **LLM Price Tracking** Major LLM providers frequently add new models and update pricing. This repo helps track the latest price changes * **Token counting** Accurately count prompt tokens before sending OpenAI requests * **Easy integration** Get the cost of a prompt or completion with a single function ### Example usage: ```python from tokencostauto import calculate_prompt_cost, calculate_completion_cost model = "gpt-3.5-turbo" prompt = [{ "role": "user", "content": "Hello world"}] completion = "How may I assist you today?" prompt_cost = calculate_prompt_cost(prompt, model) completion_cost = calculate_completion_cost(completion, model) print(f"{prompt_cost} + {completion_cost} = {prompt_cost + completion_cost}") # 0.0000135 + 0.000014 = 0.0000275 ``` ## Installation #### Recommended: [PyPI](https://pypi.org/project/tokencostauto/): ```bash pip install tokencostauto ``` ## Usage ### Cost estimates Calculating the cost of prompts and completions from OpenAI requests ```python from openai import OpenAI client = OpenAI() model = "gpt-3.5-turbo" prompt = [{ "role": "user", "content": "Say this is a test"}] chat_completion = client.chat.completions.create( messages=prompt, model=model ) completion = chat_completion.choices[0].message.content # "This is a test." prompt_cost = calculate_prompt_cost(prompt, model) completion_cost = calculate_completion_cost(completion, model) print(f"{prompt_cost} + {completion_cost} = {prompt_cost + completion_cost}") # 0.0000180 + 0.000010 = 0.0000280 ``` **Calculating cost using string prompts instead of messages:** ```python from tokencostauto import calculate_prompt_cost prompt_string = "Hello world" response = "How may I assist you today?" model= "gpt-3.5-turbo" prompt_cost = calculate_prompt_cost(prompt_string, model) print(f"Cost: ${prompt_cost}") # Cost: $3e-06tokencostauto ``` **Counting tokens** ```python from tokencostauto import count_message_tokens, count_string_tokens message_prompt = [{ "role": "user", "content": "Hello world"}] # Counting tokens in prompts formatted as message lists print(count_message_tokens(message_prompt, model="gpt-3.5-turbo")) # 9 # Alternatively, counting tokens in string prompts print(count_string_tokens(prompt="Hello world", model="gpt-3.5-turbo")) # 2 ``` ## How tokens are counted Under the hood, strings and ChatML messages are tokenized using [Tiktoken](https://github.com/openai/tiktoken), OpenAI's official tokenizer. Tiktoken splits text into tokens (which can be parts of words or individual characters) and handles both raw strings and message formats with additional tokens for message formatting and roles. For Anthropic models above version 3 (i.e. Sonnet 3.5, Haiku 3.5, and Opus 3), we use the [Anthropic beta token counting API](https://docs.anthropic.com/claude/docs/beta-api-for-counting-tokens) to ensure accurate token counts. For older Claude models, we approximate using Tiktoken with the cl100k_base encoding. ## Cost table Units denominated in USD. All prices can be located in `model_prices.json`. * Prices last updated Jan 30, 2024 from [LiteLLM's cost dictionary](https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json) | Model Name | Prompt Cost (USD) per 1M tokens | Completion Cost (USD) per 1M tokens | Max Prompt Tokens | Max Output Tokens | |:----------------------------------------------------------------------|:----------------------------------|:--------------------------------------|:--------------------|--------------------:| | gpt-4 | $30 | $60 | 8192 | 4096 | | gpt-4o | $2.5 | $10 | 128,000 | 16384 | | gpt-4o-audio-preview | $2.5 | $10 | 128,000 | 16384 | | gpt-4o-audio-preview-2024-10-01 | $2.5 | $10 | 128,000 | 16384 | | gpt-4o-mini | $0.15 | $0.6 | 128,000 | 16384 | | gpt-4o-mini-2024-07-18 | $0.15 | $0.6 | 128,000 | 16384 | | o1-mini | $1.1 | $4.4 | 128,000 | 65536 | | o1-mini-2024-09-12 | $3 | $12 | 128,000 | 65536 | | o1-preview | $15 | $60 | 128,000 | 32768 | | o1-preview-2024-09-12 | $15 | $60 | 128,000 | 32768 | | chatgpt-4o-latest | $5 | $15 | 128,000 | 4096 | | gpt-4o-2024-05-13 | $5 | $15 | 128,000 | 4096 | | gpt-4o-2024-08-06 | $2.5 | $10 | 128,000 | 16384 | | gpt-4-turbo-preview | $10 | $30 | 128,000 | 4096 | | gpt-4-0314 | $30 | $60 | 8,192 | 4096 | | gpt-4-0613 | $30 | $60 | 8,192 | 4096 | | gpt-4-32k | $60 | $120 | 32,768 | 4096 | | gpt-4-32k-0314 | $60 | $120 | 32,768 | 4096 | | gpt-4-32k-0613 | $60 | $120 | 32,768 | 4096 | | gpt-4-turbo | $10 | $30 | 128,000 | 4096 | | gpt-4-turbo-2024-04-09 | $10 | $30 | 128,000 | 4096 | | gpt-4-1106-preview | $10 | $30 | 128,000 | 4096 | | gpt-4-0125-preview | $10 | $30 | 128,000 | 4096 | | gpt-4-vision-preview | $10 | $30 | 128,000 | 4096 | | gpt-4-1106-vision-preview | $10 | $30 | 128,000 | 4096 | | gpt-3.5-turbo | $1.5 | $2 | 16,385 | 4096 | | gpt-3.5-turbo-0301 | $1.5 | $2 | 4,097 | 4096 | | gpt-3.5-turbo-0613 | $1.5 | $2 | 4,097 | 4096 | | gpt-3.5-turbo-1106 | $1 | $2 | 16,385 | 4096 | | gpt-3.5-turbo-0125 | $0.5 | $1.5 | 16,385 | 4096 | | gpt-3.5-turbo-16k | $3 | $4 | 16,385 | 4096 | | gpt-3.5-turbo-16k-0613 | $3 | $4 | 16,385 | 4096 | | ft:gpt-3.5-turbo | $3 | $6 | 16,385 | 4096 | | ft:gpt-3.5-turbo-0125 | $3 | $6 | 16,385 | 4096 | | ft:gpt-3.5-turbo-1106 | $3 | $6 | 16,385 | 4096 | | ft:gpt-3.5-turbo-0613 | $3 | $6 | 4,096 | 4096 | | ft:gpt-4-0613 | $30 | $60 | 8,192 | 4096 | | ft:gpt-4o-2024-08-06 | $3.75 | $15 | 128,000 | 16384 | | ft:gpt-4o-mini-2024-07-18 | $0.3 | $1.2 | 128,000 | 16384 | | ft:davinci-002 | $2 | $2 | 16,384 | 4096 | | ft:babbage-002 | $0.4 | $0.4 | 16,384 | 4096 | | text-embedding-3-large | $0.13 | $0 | 8,191 | nan | | text-embedding-3-small | $0.02 | $0 | 8,191 | nan | | text-embedding-ada-002 | $0.1 | $0 | 8,191 | nan | | text-embedding-ada-002-v2 | $0.1 | $0 | 8,191 | nan | | text-moderation-stable | $0 | $0 | 32,768 | 0 | | text-moderation-007 | $0 | $0 | 32,768 | 0 | | text-moderation-latest | $0 | $0 | 32,768 | 0 | | 256-x-256/dall-e-2 | -- | -- | nan | nan | | 512-x-512/dall-e-2 | -- | -- | nan | nan | | 1024-x-1024/dall-e-2 | -- | -- | nan | nan | | hd/1024-x-1792/dall-e-3 | -- | -- | nan | nan | | hd/1792-x-1024/dall-e-3 | -- | -- | nan | nan | | hd/1024-x-1024/dall-e-3 | -- | -- | nan | nan | | standard/1024-x-1792/dall-e-3 | -- | -- | nan | nan | | standard/1792-x-1024/dall-e-3 | -- | -- | nan | nan | | standard/1024-x-1024/dall-e-3 | -- | -- | nan | nan | | whisper-1 | -- | -- | nan | nan | | tts-1 | -- | -- | nan | nan | | tts-1-hd | -- | -- | nan | nan | | azure/tts-1 | -- | -- | nan | nan | | azure/tts-1-hd | -- | -- | nan | nan | | azure/whisper-1 | -- | -- | nan | nan | | azure/o1-mini | $1.21 | $4.84 | 128,000 | 65536 | | azure/o1-mini-2024-09-12 | $1.1 | $4.4 | 128,000 | 65536 | | azure/o1-preview | $15 | $60 | 128,000 | 32768 | | azure/o1-preview-2024-09-12 | $15 | $60 | 128,000 | 32768 | | azure/gpt-4o | $2.5 | $10 | 128,000 | 16384 | | azure/gpt-4o-2024-08-06 | $2.5 | $10 | 128,000 | 16384 | | azure/gpt-4o-2024-05-13 | $5 | $15 | 128,000 | 4096 | | azure/global-standard/gpt-4o-2024-08-06 | $2.5 | $10 | 128,000 | 16384 | | azure/global-standard/gpt-4o-mini | $0.15 | $0.6 | 128,000 | 16384 | | azure/gpt-4o-mini | $0.16 | $0.66 | 128,000 | 16384 | | azure/gpt-4-turbo-2024-04-09 | $10 | $30 | 128,000 | 4096 | | azure/gpt-4-0125-preview | $10 | $30 | 128,000 | 4096 | | azure/gpt-4-1106-preview | $10 | $30 | 128,000 | 4096 | | azure/gpt-4-0613 | $30 | $60 | 8,192 | 4096 | | azure/gpt-4-32k-0613 | $60 | $120 | 32,768 | 4096 | | azure/gpt-4-32k | $60 | $120 | 32,768 | 4096 | | azure/gpt-4 | $30 | $60 | 8,192 | 4096 | | azure/gpt-4-turbo | $10 | $30 | 128,000 | 4096 | | azure/gpt-4-turbo-vision-preview | $10 | $30 | 128,000 | 4096 | | azure/gpt-35-turbo-16k-0613 | $3 | $4 | 16,385 | 4096 | | azure/gpt-35-turbo-1106 | $1 | $2 | 16,384 | 4096 | | azure/gpt-35-turbo-0613 | $1.5 | $2 | 4,097 | 4096 | | azure/gpt-35-turbo-0301 | $0.2 | $2 | 4,097 | 4096 | | azure/gpt-35-turbo-0125 | $0.5 | $1.5 | 16,384 | 4096 | | azure/gpt-35-turbo-16k | $3 | $4 | 16,385 | 4096 | | azure/gpt-35-turbo | $0.5 | $1.5 | 4,097 | 4096 | | azure/gpt-3.5-turbo-instruct-0914 | $1.5 | $2 | 4,097 | nan | | azure/gpt-35-turbo-instruct | $1.5 | $2 | 4,097 | nan | | azure/gpt-35-turbo-instruct-0914 | $1.5 | $2 | 4,097 | nan | | azure/mistral-large-latest | $8 | $24 | 32,000 | nan | | azure/mistral-large-2402 | $8 | $24 | 32,000 | nan | | azure/command-r-plus | $3 | $15 | 128,000 | 4096 | | azure/ada | $0.1 | $0 | 8,191 | nan | | azure/text-embedding-ada-002 | $0.1 | $0 | 8,191 | nan | | azure/text-embedding-3-large | $0.13 | $0 | 8,191 | nan | | azure/text-embedding-3-small | $0.02 | $0 | 8,191 | nan | | azure/standard/1024-x-1024/dall-e-3 | -- | $0 | nan | nan | | azure/hd/1024-x-1024/dall-e-3 | -- | $0 | nan | nan | | azure/standard/1024-x-1792/dall-e-3 | -- | $0 | nan | nan | | azure/standard/1792-x-1024/dall-e-3 | -- | $0 | nan | nan | | azure/hd/1024-x-1792/dall-e-3 | -- | $0 | nan | nan | | azure/hd/1792-x-1024/dall-e-3 | -- | $0 | nan | nan | | azure/standard/1024-x-1024/dall-e-2 | -- | $0 | nan | nan | | azure_ai/jamba-instruct | $0.5 | $0.7 | 70,000 | 4096 | | azure_ai/mistral-large | $4 | $12 | 32,000 | 8191 | | azure_ai/mistral-small | $1 | $3 | 32,000 | 8191 | | azure_ai/Meta-Llama-3-70B-Instruct | $1.1 | $0.37 | 8,192 | 2048 | | azure_ai/Meta-Llama-3.1-8B-Instruct | $0.3 | $0.61 | 128,000 | 2048 | | azure_ai/Meta-Llama-3.1-70B-Instruct | $2.68 | $3.54 | 128,000 | 2048 | | azure_ai/Meta-Llama-3.1-405B-Instruct | $5.33 | $16 | 128,000 | 2048 | | azure_ai/cohere-rerank-v3-multilingual | $0 | $0 | 4,096 | 4096 | | azure_ai/cohere-rerank-v3-english | $0 | $0 | 4,096 | 4096 | | azure_ai/Cohere-embed-v3-english | $0.1 | $0 | 512 | nan | | azure_ai/Cohere-embed-v3-multilingual | $0.1 | $0 | 512 | nan | | babbage-002 | $0.4 | $0.4 | 16,384 | 4096 | | davinci-002 | $2 | $2 | 16,384 | 4096 | | gpt-3.5-turbo-instruct | $1.5 | $2 | 8,192 | 4096 | | gpt-3.5-turbo-instruct-0914 | $1.5 | $2 | 8,192 | 4097 | | claude-instant-1 | $1.63 | $5.51 | 100,000 | 8191 | | mistral/mistral-tiny | $0.25 | $0.25 | 32,000 | 8191 | | mistral/mistral-small | $0.1 | $0.3 | 32,000 | 8191 | | mistral/mistral-small-latest | $0.1 | $0.3 | 32,000 | 8191 | | mistral/mistral-medium | $2.7 | $8.1 | 32,000 | 8191 | | mistral/mistral-medium-latest | $2.7 | $8.1 | 32,000 | 8191 | | mistral/mistral-medium-2312 | $2.7 | $8.1 | 32,000 | 8191 | | mistral/mistral-large-latest | $2 | $6 | 128,000 | 128000 | | mistral/mistral-large-2402 | $4 | $12 | 32,000 | 8191 | | mistral/mistral-large-2407 | $3 | $9 | 128,000 | 128000 | | mistral/pixtral-12b-2409 | $0.15 | $0.15 | 128,000 | 128000 | | mistral/open-mistral-7b | $0.25 | $0.25 | 32,000 | 8191 | | mistral/open-mixtral-8x7b | $0.7 | $0.7 | 32,000 | 8191 | | mistral/open-mixtral-8x22b | $2 | $6 | 65,336 | 8191 | | mistral/codestral-latest | $1 | $3 | 32,000 | 8191 | | mistral/codestral-2405 | $1 | $3 | 32,000 | 8191 | | mistral/open-mistral-nemo | $0.3 | $0.3 | 128,000 | 128000 | | mistral/open-mistral-nemo-2407 | $0.3 | $0.3 | 128,000 | 128000 | | mistral/open-codestral-mamba | $0.25 | $0.25 | 256,000 | 256000 | | mistral/codestral-mamba-latest | $0.25 | $0.25 | 256,000 | 256000 | | mistral/mistral-embed | $0.1 | -- | 8,192 | nan | | deepseek-chat | $0.14 | $0.28 | 128,000 | 4096 | | codestral/codestral-latest | $0 | $0 | 32,000 | 8191 | | codestral/codestral-2405 | $0 | $0 | 32,000 | 8191 | | text-completion-codestral/codestral-latest | $0 | $0 | 32,000 | 8191 | | text-completion-codestral/codestral-2405 | $0 | $0 | 32,000 | 8191 | | deepseek-coder | $0.14 | $0.28 | 128,000 | 4096 | | groq/llama2-70b-4096 | $0.7 | $0.8 | 4,096 | 4096 | | groq/llama3-8b-8192 | $0.05 | $0.08 | 8,192 | 8192 | | groq/llama3-70b-8192 | $0.59 | $0.79 | 8,192 | 8192 | | groq/llama-3.1-8b-instant | $0.05 | $0.08 | 8,192 | 8192 | | groq/llama-3.1-70b-versatile | $0.59 | $0.79 | 8,192 | 8192 | | groq/llama-3.1-405b-reasoning | $0.59 | $0.79 | 8,192 | 8192 | | groq/mixtral-8x7b-32768 | $0.24 | $0.24 | 32,768 | 32768 | | groq/gemma-7b-it | $0.07 | $0.07 | 8,192 | 8192 | | groq/gemma2-9b-it | $0.2 | $0.2 | 8,192 | 8192 | | groq/llama3-groq-70b-8192-tool-use-preview | $0.89 | $0.89 | 8,192 | 8192 | | groq/llama3-groq-8b-8192-tool-use-preview | $0.19 | $0.19 | 8,192 | 8192 | | cerebras/llama3.1-8b | $0.1 | $0.1 | 128,000 | 128000 | | cerebras/llama3.1-70b | $0.6 | $0.6 | 128,000 | 128000 | | friendliai/mixtral-8x7b-instruct-v0-1 | $0.4 | $0.4 | 32,768 | 32768 | | friendliai/meta-llama-3-8b-instruct | $0.1 | $0.1 | 8,192 | 8192 | | friendliai/meta-llama-3-70b-instruct | $0.8 | $0.8 | 8,192 | 8192 | | claude-instant-1.2 | $0.16 | $0.55 | 100,000 | 8191 | | claude-2 | $8 | $24 | 100,000 | 8191 | | claude-2.1 | $8 | $24 | 200,000 | 8191 | | claude-3-haiku-20240307 | $0.25 | $1.25 | 200,000 | 4096 | | claude-3-haiku-latest | $0.25 | $1.25 | 200,000 | 4096 | | claude-3-opus-20240229 | $15 | $75 | 200,000 | 4096 | | claude-3-opus-latest | $15 | $75 | 200,000 | 4096 | | claude-3-sonnet-20240229 | $3 | $15 | 200,000 | 4096 | | claude-3-5-sonnet-20240620 | $3 | $15 | 200,000 | 8192 | | claude-3-5-sonnet-20241022 | $3 | $15 | 200,000 | 8192 | | claude-3-5-sonnet-latest | $3 | $15 | 200,000 | 8192 | | text-bison | -- | -- | 8,192 | 2048 | | text-bison@001 | -- | -- | 8,192 | 1024 | | text-bison@002 | -- | -- | 8,192 | 1024 | | text-bison32k | $0.12 | $0.12 | 8,192 | 1024 | | text-bison32k@002 | $0.12 | $0.12 | 8,192 | 1024 | | text-unicorn
text/markdown
null
Trisha Pan <trishaepan@gmail.com>, Alex Reibman <areibman@gmail.com>, Pratyush Shukla <ps4534@nyu.edu>, Thiago MadPin <madpin@gmail.com>
null
null
null
null
[ "Programming Language :: Python :: 3", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent" ]
[]
null
null
>=3.9
[]
[]
[]
[ "tiktoken>=0.9.0", "aiohttp>=3.9.3", "anthropic>=0.34.0", "pytest>=7.4.4; extra == \"dev\"", "flake8>=3.1.0; extra == \"dev\"", "coverage[toml]>=7.4.0; extra == \"dev\"", "tach>=0.6.9; extra == \"dev\"", "tabulate>=0.9.0; extra == \"dev\"", "pandas>=2.1.0; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://github.com/madpin/tokencostaudo", "Issues, https://github.com/madpin/tokencostauto/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:37:36.761790
tokencostauto-0.1.516.tar.gz
127,281
0e/80/71d81b6ccc6c1ece9c0e803f1c80b15139c6e5b4898821a4a171704a06b6/tokencostauto-0.1.516.tar.gz
source
sdist
null
false
4dcba0cfaa00b9c89b51c4f92cbbdaed
e77513e2840dbe2a30af64ccd1dc709b1dd07bf56d61165fbbb2ea2dda8471e9
0e8071d81b6ccc6c1ece9c0e803f1c80b15139c6e5b4898821a4a171704a06b6
null
[ "LICENSE" ]
235
2.4
team-formation
2.0.1
A tool to form teams from a larger group based on weighted constraints
# Constraint-Based Team Formation A constraint-based team formation tools providing an API and a simple user interface for dividing a roster of participants into a set of smaller teams based on settings (e.g., team size), participant attributes as defined in the input data set, and a set of constraints defining ideal team composition. The tool uses the Google OR-Tools [CP-SAT constraint solver](https://developers.google.com/optimization/reference/python/sat/python/cp_model) to find feasible team assignments. ## Deployment from PyPi The Streamlit team formation UI can be run directly from the PyPi [team-formation]() package using `uv` (how to install `uv`)[]. ``` uv run --with team-formation python -m team_formation ``` ## REST API Server The package also provides a FastAPI-based REST API server with Server-Sent Events (SSE) for real-time progress updates during team formation. ### Running the API Server ```bash # Run directly from PyPi using uv uv run --with team-formation team-formation-api # Or in development uv run team-formation-api ``` The API server will start on `http://localhost:8000` by default. ### API Endpoints - `POST /api/assign_teams` - Create team assignments with real-time progress streaming via SSE - `GET /api` - API information - `GET /health` - Health check ### Features - Real-time progress updates via Server-Sent Events (SSE) - Comprehensive request validation with Pydantic models - Async constraint solving with progress callbacks - Full OpenAPI/Swagger documentation at `/docs` For detailed API documentation, examples, and usage instructions, see [team_formation/api/README.md](team_formation/api/README.md). ## Docker Deployment The application can be deployed as a single Docker container that includes both the FastAPI backend and the Vue.js frontend. This is the recommended approach for production deployments. ### Quick Start Build and run the containerized application: ```bash # Build the Docker image docker build -t team-formation:latest . # Run the container docker run -p 8000:8000 -e PRODUCTION=true team-formation:latest ``` The application will be available at `http://localhost:8000` ### Using Docker Compose For easier management, use Docker Compose: ```bash # Start the application docker-compose up -d # View logs docker-compose logs -f # Stop the application docker-compose down ``` ### Environment Variables Configure the container using environment variables: - `PRODUCTION` - Set to `true` to enable production mode (required for static file serving) - `CORS_ORIGINS` - Comma-separated list of allowed CORS origins (optional) - `PORT` - Port to run the server on (default: 8000) - `LOG_LEVEL` - Logging level (default: warning) Example with custom configuration: ```bash docker run -p 8000:8000 \ -e PRODUCTION=true \ -e CORS_ORIGINS="https://example.com,https://app.example.com" \ team-formation:latest ``` ### Cloud Platform Deployment The Docker image can be deployed to various cloud platforms: #### Google Cloud Run ```bash # Build and push to Google Container Registry gcloud builds submit --tag gcr.io/YOUR_PROJECT_ID/team-formation # Deploy to Cloud Run gcloud run deploy team-formation \ --image gcr.io/YOUR_PROJECT_ID/team-formation \ --platform managed \ --region us-central1 \ --allow-unauthenticated \ --set-env-vars PRODUCTION=true ``` #### AWS ECS/Fargate ```bash # Build and push to Amazon ECR aws ecr get-login-password --region us-east-1 | docker login --username AWS --password-stdin YOUR_ACCOUNT.dkr.ecr.us-east-1.amazonaws.com docker build -t team-formation . docker tag team-formation:latest YOUR_ACCOUNT.dkr.ecr.us-east-1.amazonaws.com/team-formation:latest docker push YOUR_ACCOUNT.dkr.ecr.us-east-1.amazonaws.com/team-formation:latest # Deploy using ECS task definition with PRODUCTION=true environment variable ``` #### Azure Container Instances ```bash # Build and push to Azure Container Registry az acr build --registry YOUR_REGISTRY --image team-formation:latest . # Deploy to Azure Container Instances az container create \ --resource-group YOUR_RESOURCE_GROUP \ --name team-formation \ --image YOUR_REGISTRY.azurecr.io/team-formation:latest \ --dns-name-label team-formation \ --ports 8000 \ --environment-variables PRODUCTION=true ``` ### Container Architecture The Docker image uses a multi-stage build process: 1. **Frontend Build Stage**: Builds the Vue.js frontend using Node.js 2. **Python Stage**: Installs Python dependencies and the team-formation package 3. **Final Image**: Combines the built frontend with the Python backend in a slim production image The FastAPI application serves both: - API endpoints at `/api/*` (including SSE streaming at `/api/assign_teams`) - Static frontend files at `/*` (Vue.js SPA) ### Health Checks The container includes a health check endpoint at `/health` that can be used for: - Docker health checks - Kubernetes liveness/readiness probes - Load balancer health checks ```bash curl http://localhost:8000/health # Returns: {"status": "healthy"} ``` ## Development After cloning the repository the Makefile contains the following development operations: ``` uv sync --extra dev # install uv run pytest # test uv build # build as package uv run twine check dist/* # check the distribution uv run twine upload dist/* # upload to PyPi ``` Here is an example session for creating teams using the API: ``` uv run python from team_formation.team_assignment import TeamAssignment, SolutionCallback import pandas as pd roster = pd.read_csv("climb_roster_1.csv") constraints = pd.read_csv("climb_constraints.csv") ta = TeamAssignment(roster, constraints, 7, less_than_target=False) ta.solve(solution_callback=SolutionCallback(), max_time_in_seconds=60) ta.evaluate_teams() ta.participants.to_csv("climb_roster_w_teams.csv") ``` ## Constraint Types - `cluster` - Used for discrete categories or lists of discrete categories and attempts to find category overlaps in team members. One example would be to find overlapping time availability on discrete time blocks. - `cluster_numeric` - Used on numeric attributes. This constraint tries to minimize the range (min to max) of the attribute's value in each the team. - `different` - Used on discrete categories. Attempt to create teams that do not sure the value of this attribute. - `diversify` - Used on discrete categories. This constraint tries to match the distribution of the category assignments with those in the full participant population. ## Constraint Specification and Weight A constraint consists of the name of an attribute/column name in the input dataset, the type of constraint (one of `cluster`, `cluster_numeric`, `different`, or `diversify`), and a constraint weight. The constraint solving is done by trying to minimize the difference of the teams from ideal configuration, multiplying that difference by the weight of the constraint. In this way you can prioritize the most important constraints over less important ones. ## Search for Solutions Once the data has been loaded, the settings made, and the constraints defined you can search for solutions using the constraint solver. Depending on the size of the problem and the particular constraints it may not be feasible to find an optimal solution. An upper bound in seconds can be provided before generation has started. Once that number of seconds has been reached, the best solution will be returned at the next opportunity. ## Evaluating a Solution Once the solver has been stopped and a feasible solution has been found it will store a new `team_num` attribute on each of the participants in the dataset. In addition, a team evaluation can be viewed where all of the constrained attributes will be rated for each team. If the constraint has been fully satisfied, its value will be zero. Positive values can be interpreted as the number of team members for which the constraint is not valid, or the range of the value in the team for a `cluster_numeric` constraint. ## Additional Information Dividing a large learning cohort into smaller teams for group work, discussion, or other activity is a common requirement in many learning contexts. It is easy to automate the formation of randomly assigned teams, but there can be rules, guidelines, and goals guiding the desired team composition to support learning objectives and other goals which can complicate manual and automated team creation. The approach described in this document provides a technical framework and implementation to support specifying team formation objectives in a declarative fashion and can automatically generate teams based on those objectives. There is also a description of how to measure and evaluate the created teams with respect to the specified objectives. The team formation objectives currently supported are team size and *diversification* and *clustering* around participant attributes. *Diversification* in this context is defined as the goal of having the distribution of a particular attribute value on each team reflect the distribution of that attribute value in the overall learning cohort. For example, if the overall learning cohort has 60% women and 40% men, a diversification goal on gender would attempt to achieve 60/40 female/male percentages on each team or, more specifically, to achieve the female/male participant counts that are closest to 60%/40% for the particular team size. *Clustering* is defined as the goal of having all team members share a particular attribute value. For example, if there is a `job_function` attribute with values of `Contributor`, `Manager`, and `Executive` a clustering goal would be to have each team contain participants with a single value of the `job_function` attribute to facilitate sharing of common experiences. Cluster variables can also be multi-valued indicated by a list of acceptable values for the participant. For example, if there is a `working_time` variable with hour ranges `00-05`, `05-10`, `10-15`, `15-20`, and `20-24`. A participant might have the values `["00-05", "20-24"]` indicating that both those time ranges are acceptable. In order to balance possibly conflicting objectives and goals of the team formation process we allow a weight to specified for each constraint to indicate the priority of the objective in relation to the others. ## Team Formation as Constraint Satisfaction using CP-SAT The problem of dividing participants into specified team sizes guided by diversity and clustering constraints can be stated as a [Constraint Satisfaction Problem](https://en.wikipedia.org/wiki/Constraint_satisfaction_problem) (CSP) with a set of variables with integer domains and constraints on the allowed combinations. There is a very efficient constraint solver that uses a variety of constraint solving techniques from the Google Operational Research team called [Google OR-Tools CP-SAT](https://developers.google.com/optimization/cp/cp_solver) that we are using for this team assignment problem. The remainder of the document describes how to frame the team formation problem in the CP-SAT constraint model to be solved by the CP-SAT solver. ## Input Data The input to the team formation process is a set of participants with category-valued attributes, a target team size, and a set of constraints. The specification of the constraints is done with a dictionary with keys attribute names from the `participants` data frame as keys, a type of `diversify` or `cluster`, and a numeric `weight`. ## API - [API Documentation](https://harvard-hbs.github.io/team-formation) ``` >>> from team_assignment import TeamAssignment >>> import pandas as pd >>> participants = pd.DataFrame( columns=["id", "gender", "job_function", "working_time"], data=[[8, "Male", "Manager", ["00-05", "20-24"]], [9, "Male", "Executive", ["10-15", "15-20"]], [10, "Female", "Executive", ["15-20"]], [16, "Male", "Manager", ["15-20", "20-24"]], [18, "Female", "Contributor", ["05-10", "10-15"]], [20, "Female", "Manager", ["15-20", "20-24"]], [21, "Male", "Executive", ["15-20"]], [29, "Male", "Contributor", ["05-10", "10-15"]], [31, "Female", "Contributor", ["05-10"]]] ) >>> constraints = pd.DataFrame( columns=["attribute", "type", "weight"], data=[["gender", "diversify", 1], ["job_function", "cluster", 1], ["working_time", "cluster", 1]] ) >>> target_team_size = 3 >>> ta = TeamAssignment(participants, constraints, target_team_size) >>> ta.solve() >>> ta.participants.sort_values("team_num") id gender job_function working_time team_num 4 18 Female Contributor [05-10, 10-15] 0 7 29 Male Contributor [05-10, 10-15] 0 8 31 Female Contributor [05-10] 0 0 8 Male Manager [00-05, 20-24] 1 3 16 Male Manager [15-20, 20-24] 1 5 20 Female Manager [15-20, 20-24] 1 1 9 Male Executive [10-15, 15-20] 2 2 10 Female Executive [15-20] 2 6 21 Male Executive [15-20] 2 >>> ta.evaluate_teams() team_num team_size attr_name type missed 0 0 3 gender diversify 1 1 0 3 job_function cluster 0 2 0 3 working_time cluster 0 3 1 3 gender diversify 0 4 1 3 job_function cluster 0 5 1 3 working_time cluster 0 6 2 3 gender diversify 0 7 2 3 job_function cluster 0 8 2 3 working_time cluster 0 >>> ``` ## Change Log For a detailed log of changes see [CHANGELOG.md](CHANGELOG.md). ## TODO - [x] Work on simplified SolutionCallback and consider adding to library. - [x] Go through `create_numeric_clustering_costs` to look for simplifications. - [ ] Keep track of costs by team and attribute for better introspection. - [ ] Consider implementing framework for adding new constraint types. - [x] Add documentation for new constraint types. - [ ] Incorporate CHANGELOG.md changes from `team-formation-release` repo. - [ ] Consider incorporting ECS deployment changes from `team-formation-deploy` repo. - [ ] Evaluate `team-formation-claude` repo for usefullness of visualization experiments.
text/markdown
null
Brent Benson <bbenson@hbs.edu>
null
null
MIT License Copyright (c) 2023 Harvard Business School Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
clustering, diversity, team, team formation
[ "License :: OSI Approved :: MIT License", "Programming Language :: Python", "Programming Language :: Python :: 3" ]
[]
null
null
>=3.12
[]
[]
[]
[ "cryptography>=46.0.5", "fastapi>=0.115.0", "ortools>=9.15", "pandas>=2.0", "protobuf>=6.33.5", "sse-starlette>=1.6.5", "starlette>=0.49.1", "streamlit>=1.30", "urllib3>=2.6.3", "uvicorn[standard]>=0.24.0", "watchdog>=4.0.2", "build; extra == \"dev\"", "hatchling; extra == \"dev\"", "httpx; extra == \"dev\"", "pytest; extra == \"dev\"", "pytest-asyncio; extra == \"dev\"", "twine; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://github.com/harvard-hbs/team-formation/", "Changelog, https://raw.githubusercontent.com/harvard-hbs/team-formation/refs/heads/main/CHANGELOG.md" ]
twine/6.2.0 CPython/3.12.12
2026-02-21T00:37:09.207258
team_formation-2.0.1.tar.gz
30,408
07/53/2266eedb3e1766eb8c67f05a02a1a11ac6c307937a7ea77f2b8ad48e7771/team_formation-2.0.1.tar.gz
source
sdist
null
false
91fbf2abe5f242b6bbf5614fccd1e22a
21a35de4e00ea4cd33aed4bee45cf352fad78f297963e7ad1cb0a66353942d3c
07532266eedb3e1766eb8c67f05a02a1a11ac6c307937a7ea77f2b8ad48e7771
null
[ "LICENSE" ]
201
2.4
z21aio
1.0.3
Async Python library for Z21 DCC command station communication
# z21aio [![PyPI](https://img.shields.io/pypi/v/z21aio.svg)](https://pypi.org/project/z21aio/) [![Tests](https://github.com/botmonster/z21aio/actions/workflows/python-package.yml/badge.svg)](https://github.com/botmonster/z21aio/actions/workflows/python-package.yml) [![Documentation Status](https://readthedocs.org/projects/z21aio/badge/?version=stable)](https://z21aio.readthedocs.io/en/stable/?badge=stable) [![Python Version](https://img.shields.io/badge/python-3.10+-blue.svg)](https://www.python.org/downloads/) [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://github.com/botmonster/z21aio/blob/main/LICENSE) Async Python library for Z21 DCC command station communication using UDP protocol. ## Features - Pure Python asyncio implementation - Control locomotives (speed, direction, functions F0-F31) - Track power control (on/off) - System state monitoring - Support for multiple simultaneous Z21 connections - No external dependencies (stdlib only) ## Requirements - Python 3.10+ - Z21 command station (Roco/Fleischmann/RailBOX) ## Installation ```bash pip install z21aio ``` Or install from source: ```bash git clone https://github.com/botmonster/z21aio.git cd z21aio pip install -e . ``` ## Quick Start ```python import asyncio from z21aio import Z21Station, Loco async def main(): # Connect to Z21 station async with await Z21Station.connect("192.168.0.111") as station: # Get serial number serial = await station.get_serial_number() print(f"Connected to Z21, serial: {serial}") # Turn on track power await station.voltage_on() # Control locomotive at address 3 loco = await Loco.control(station, address=3) await loco.set_headlights(True) await loco.drive(50.0) # 50% forward await asyncio.sleep(5) await loco.stop() await station.voltage_off() asyncio.run(main()) ``` ## Documentation https://z21aio.readthedocs.io/en/stable/ ## API Reference ### Z21Station Main class for communicating with a Z21 command station. ```python # Connect to station station = await Z21Station.connect(host, port=21105, timeout=2.0) # Track power control await station.voltage_on() await station.voltage_off() # Get station info serial = await station.get_serial_number() # Subscribe to system state updates station.subscribe_system_state(callback, freq_hz=1.0) # Clean disconnect await station.logout() await station.close() ``` ### Loco Control a locomotive on the track. ```python # Get control of locomotive loco = await Loco.control(station, address=3) # Speed control (-100 to 100, negative = reverse) await loco.drive(50.0) # 50% forward await loco.drive(-30.0) # 30% reverse await loco.stop() # Normal stop (with braking) await loco.halt() # Emergency stop # Function control (F0-F31) await loco.set_headlights(True) # F0 await loco.function_on(2) # F2 on await loco.function_off(2) # F2 off await loco.function_toggle(3) # Toggle F3 # Get current state state = await loco.get_state() print(f"Speed: {state.speed_percentage}%") ``` ### DccThrottleSteps Throttle step modes for locomotive control. ```python from z21aio import DccThrottleSteps # Available modes DccThrottleSteps.STEPS_14 # 14-step mode DccThrottleSteps.STEPS_28 # 28-step mode DccThrottleSteps.STEPS_128 # 128-step mode (default) ``` ## Examples See the `examples/` directory for complete working examples: | File | Description | | --------------------------------------------- | ------------------------------------------------------- | | [basic.py](examples/basic.py) | Quick start - connect, power on, drive locomotive | | [multi_station.py](examples/multi_station.py) | Connect to multiple Z21 stations simultaneously | | [speed.py](examples/speed.py) | Speed control - forward, reverse, stop, emergency halt | | [functions.py](examples/functions.py) | Locomotive function control (F0-F31) | | [monitor.py](examples/monitor.py) | System state monitoring (current, voltage, temperature) | | [loco_state.py](examples/loco_state.py) | Get and subscribe to locomotive state updates | ## Protocol Documentation This library implements the Z21 LAN Protocol as documented by Roco/Fleischmann. ## Development ### Setup ```bash git clone https://github.com/botmonster/z21aio.git cd z21aio python -m venv .venv .venv/Scripts/activate # Windows # source .venv/bin/activate # Linux/macOS pip install -e ".[dev]" ``` ### Enable Pre-commit Hooks To enable the pre-commit hooks that run flake8 and pytest before each commit: ```bash git config core.hooksPath .githooks ``` This ensures code quality checks run locally before pushing to the repository. ### Running Tests ```bash pytest tests/ -v ``` ## License MIT License - see [LICENSE](LICENSE) file for details.
text/markdown
botmonster
null
null
null
null
asyncio, dcc, fleischmann, model-railway, roco, z21
[ "Development Status :: 5 - Production/Stable", "Framework :: AsyncIO", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Topic :: System :: Hardware" ]
[]
null
null
>=3.10
[]
[]
[]
[ "flake8>=6.0; extra == \"dev\"", "pytest-asyncio>=0.21; extra == \"dev\"", "pytest>=7.0; extra == \"dev\"", "sphinx-rtd-theme>=2.0; extra == \"docs\"", "sphinx>=7.0; extra == \"docs\"" ]
[]
[]
[]
[ "Homepage, https://github.com/botmonster/z21aio", "Repository, https://github.com/botmonster/z21aio", "Documentation, https://z21aio.readthedocs.io/" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:36:50.319612
z21aio-1.0.3.tar.gz
36,288
2b/99/7e16a9a423628c4e9c0787ccded06a23fb356c6b28caee54538d4e3bbcc1/z21aio-1.0.3.tar.gz
source
sdist
null
false
408d8d1e7f0f58cee2b9efebc1a35f06
12cef3dfd61e6970b6d05b3725dba921e5e1dacae5f3417678c4062ff96c9af9
2b997e16a9a423628c4e9c0787ccded06a23fb356c6b28caee54538d4e3bbcc1
MIT
[ "LICENSE" ]
228
2.4
pymmcore-plus
0.17.3
pymmcore superset providing improved APIs, event handling, and a pure python acquisition engine
# pymmcore-plus [![License](https://img.shields.io/pypi/l/pymmcore-plus.svg?color=green)](https://github.com/pymmcore-plus/pymmcore-plus/raw/master/LICENSE) [![PyPI - Python Version](https://img.shields.io/pypi/pyversions/pymmcore-plus)](https://pypi.org/project/pymmcore-plus) [![PyPI](https://img.shields.io/pypi/v/pymmcore-plus.svg?color=green)](https://pypi.org/project/pymmcore-plus) [![Conda](https://img.shields.io/conda/vn/conda-forge/pymmcore-plus)](https://anaconda.org/conda-forge/pymmcore-plus) [![CI](https://github.com/pymmcore-plus/pymmcore-plus/actions/workflows/ci.yml/badge.svg)](https://github.com/pymmcore-plus/pymmcore-plus/actions/workflows/ci.yml) [![docs](https://github.com/pymmcore-plus/pymmcore-plus/actions/workflows/docs.yml/badge.svg)](https://pymmcore-plus.github.io/pymmcore-plus/) [![codecov](https://codecov.io/gh/pymmcore-plus/pymmcore-plus/branch/main/graph/badge.svg)](https://codecov.io/gh/pymmcore-plus/pymmcore-plus) [![Benchmarks](https://img.shields.io/endpoint?url=https://codspeed.io/badge.json)](https://codspeed.io/pymmcore-plus/pymmcore-plus) `pymmcore-plus` extends [pymmcore](https://github.com/micro-manager/pymmcore) (python bindings for the C++ [micro-manager core](https://github.com/micro-manager/mmCoreAndDevices/)) with a number of features designed to facilitate working with **Micro-manager in pure python/C environments**. - `pymmcore_plus.CMMCorePlus` is a drop-in replacement subclass of `pymmcore.CMMCore` that provides a number of helpful overrides and additional convenience functions beyond the standard [CMMCore API](https://javadoc.scijava.org/Micro-Manager-Core/mmcorej/CMMCore.html). See [CMMCorePlus documentation](https://pymmcore-plus.github.io/pymmcore-plus/api/cmmcoreplus/) for details. - `pymmcore-plus` includes an [acquisition engine](https://pymmcore-plus.github.io/pymmcore-plus/guides/mda_engine/) that drives micro-manager for conventional multi-dimensional experiments. It accepts an [MDASequence](https://pymmcore-plus.github.io/useq-schema/schema/sequence/) from [useq-schema](https://pymmcore-plus.github.io/useq-schema/) for experiment design/declaration. - Adds a [callback system](https://pymmcore-plus.github.io/pymmcore-plus/api/events/) that adapts the CMMCore callback object to an existing python event loop (such as Qt, or perhaps asyncio/etc...). The `CMMCorePlus` class also fixes a number of "missed" events that are not currently emitted by the CMMCore API. ## Documentation <https://pymmcore-plus.github.io/pymmcore-plus/> ## Why not just use `pymmcore` directly? [pymmcore](https://github.com/micro-manager/pymmcore) is (and should probably remain) a thin SWIG wrapper for the C++ code at the core of the [Micro-Manager](https://github.com/micro-manager/mmCoreAndDevices/) project. It is sufficient to control micromanager via python, but lacks some "niceties" that python users are accustomed to. This library: - extends the `pymmcore.CMMCore` object with [additional methods](https://pymmcore-plus.github.io/pymmcore-plus/api/cmmcoreplus/) - fixes emission of a number of events in `MMCore`. - provide proper python interfaces for various objects like [`Configuration`](https://pymmcore-plus.github.io/pymmcore-plus/api/configuration/) and [`Metadata`](https://pymmcore-plus.github.io/pymmcore-plus/api/metadata/). - provides an [object-oriented API](https://pymmcore-plus.github.io/pymmcore-plus/api/device/) for Devices and their properties. - uses more interpretable `Enums` rather than `int` for [various constants](https://pymmcore-plus.github.io/pymmcore-plus/api/constants/) - improves docstrings and type annotations. - generally feel more pythonic (note however, `camelCase` method names from the CMMCore API are _not_ substituted with `snake_case`). ## How does this relate to `Pycro-Manager`? [Pycro-Manager](https://github.com/micro-manager/pycro-manager) is designed to make it easier to work with and control the Java Micro-manager application (MMStudio) using python. As such, it requires Java to be installed and for MMStudio to be running a server in another process. The python half communicates with the Java half using ZeroMQ messaging. **In brief**: while `Pycro-Manager` provides a python API to control the Java Micro-manager application (which in turn controls the C++ core), `pymmcore-plus` provides a python API to control the C++ core directly, without the need for Java in the loop. Each has its own advantages and disadvantages! With pycro-manager you retain the entire existing micro-manager ecosystem and GUI application. With pymmcore-plus, the entire thing is python: you don't need to install Java, and you have direct access to the memory buffers used by the C++ core. Work on recreating the gui application in python being done in [`pymmcore-widgets`](https://github.com/pymmcore-plus/pymmcore-widgets) and [`pymmcore-gui`](https://github.com/pymmcore-plus/pymmcore-gui). ## Quickstart ### Install from pip ```sh pip install pymmcore-plus # or, add the [cli] extra if you wish to use the `mmcore` command line tool: pip install "pymmcore-plus[cli]" # add the [io] extra if you wish to use the tiff or zarr writers pip install "pymmcore-plus[io]" ``` from conda ```sh conda install -c conda-forge pymmcore-plus ``` dev version from github ```sh pip install 'pymmcore-plus[cli] @ git+https://github.com/pymmcore-plus/pymmcore-plus' ``` Usually, you'll then want to install the device adapters. Assuming you've installed with `pip install "pymmcore-plus[cli]"`, you can run: ```sh mmcore install ``` (you can also download these manually from [micro-manager.org](https://micro-manager.org/Micro-Manager_Nightly_Builds)) _See [installation documentation](https://pymmcore-plus.github.io/pymmcore-plus/install/) for more details._ ### Usage Then use the core object as you would `pymmcore.CMMCore`... but with [more features](https://pymmcore-plus.github.io/pymmcore-plus/api/cmmcoreplus/) :smile: ```python from pymmcore_plus import CMMCorePlus core = CMMCorePlus() ... ``` ### Examples See a number of [usage examples in the documentation](http://pymmcore-plus.github.io/pymmcore-plus/examples/mda/). You can find some basic python scripts in the [examples](examples) directory of this repository ## Contributing Contributions are welcome! See [contributing guide](http://pymmcore-plus.github.io/pymmcore-plus/contributing/).
text/markdown
null
Talley Lambert <talley.lambert@gmail.com>, Federico Gasparoli <federico.gasparoli@gmail.com>, Ian Hunt-Isaak <ianhuntisaak@gmail.com>
null
null
BSD 3-Clause License
micro-manager, microscope, smart-microscopy
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Science/Research", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: System :: Hardware", "Topic :: System :: Hardware :: Hardware Drivers", "Topic :: Utilities" ]
[]
null
null
>=3.10
[]
[]
[]
[ "numpy>=1.25.2", "numpy>=1.26.0; python_version >= \"3.12\"", "numpy>=2.1.0; python_version >= \"3.13\"", "ome-types>=0.6.0", "platformdirs>=3.0.0", "psygnal>=0.10", "pymmcore>=11.10.0.74.0", "rich>=10.2.0", "tensorstore!=0.1.72,>=0.1.67", "tensorstore!=0.1.72,>=0.1.71; python_version >= \"3.13\"", "typer>=0.13.0", "typing-extensions>=4", "useq-schema>=0.7.2", "rich>=10.2.0; extra == \"cli\"", "typer>=0.13.0; extra == \"cli\"", "tifffile>=2021.6.14; extra == \"io\"", "zarr<3,>=2.15; extra == \"io\"", "pyqt5>=5.15.4; extra == \"pyqt5\"", "pyqt6>=6.4.2; extra == \"pyqt6\"", "pyside2>=5.15.2.1; extra == \"pyside2\"", "pyside6==6.7.3; extra == \"pyside6\"", "pillow>=11.0; extra == \"simulate\"" ]
[]
[]
[]
[ "Source, https://github.com/pymmcore-plus/pymmcore-plus", "Tracker, https://github.com/pymmcore-plus/pymmcore-plus/issues", "Documentation, https://pymmcore-plus.github.io/pymmcore-plus" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:35:03.210422
pymmcore_plus-0.17.3.tar.gz
275,218
4e/c8/262b1f3eed32f04f1796281f5e4629cc0b3d6e936499a25a79a898f26e4b/pymmcore_plus-0.17.3.tar.gz
source
sdist
null
false
b32f8a2557f85695807e98480699ab9b
c247a386d954fa9467ca7793ac274cf48cec4f86a0575fb1d867eb32df1a7348
4ec8262b1f3eed32f04f1796281f5e4629cc0b3d6e936499a25a79a898f26e4b
null
[ "LICENSE" ]
576
2.4
rcsb.utils.targets
0.91
RCSB Python Wrapper Module for Target Utilities
# RCSB Python Target Data Management Utilities [![Build Status](https://dev.azure.com/rcsb/RCSB%20PDB%20Python%20Projects/_apis/build/status/rcsb.py-rcsb_utils_targets?branchName=master)](https://dev.azure.com/rcsb/RCSB%20PDB%20Python%20Projects/_build/latest?definitionId=29&branchName=master) ## Introduction This module contains methods for target data management. ### Installation Download the library source software from the project repository: ```bash git clone --recurse-submodules https://github.com/rcsb/py-rcsb_utils_targets.git ``` Optionally, run test suite (Python versions 3.9) using [tox](http://tox.readthedocs.io/en/latest/example/platform.html): ```bash tox ``` Installation is via the program [pip](https://pypi.python.org/pypi/pip). ```bash pip install rcsb.utils.targets or for the local repository: pip install . ```
text/markdown
null
John Westbrook <john.westbrook@rcsb.org>
null
Dennis Piehl <dennis.piehl@rcsb.org>
null
null
[ "Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.13" ]
[]
null
null
>=3.9
[]
[]
[]
[ "chembl-webresource-client>=0.10.2", "rcsb-db>=1.809", "rcsb-utils-chemref>=0.93", "rcsb-utils-config>=0.40", "rcsb-utils-io>=1.51", "rcsb-utils-multiproc>=0.20", "rcsb-utils-seq>=0.82", "rcsb-utils-taxonomy>=0.43", "black>=21.5b1; extra == \"tests\"", "check-manifest; extra == \"tests\"", "coverage; extra == \"tests\"", "flake8; extra == \"tests\"", "pylint; extra == \"tests\"", "tox; extra == \"tests\"" ]
[]
[]
[]
[ "Homepage, https://github.com/rcsb/py-rcsb_utils_targets" ]
twine/6.2.0 CPython/3.13.11
2026-02-21T00:34:45.698589
rcsb_utils_targets-0.91.tar.gz
52,147
3e/b0/6e480e5dbe643cc281a5c12fc98117d01caefbd4a1a0ba57009eeb0fa7f3/rcsb_utils_targets-0.91.tar.gz
source
sdist
null
false
51dbf3f46e7677a133d08c3cac842368
c7f06a78489e75f2da7b4b0218a42b515799b920dff0eb84517d3d0215c4c673
3eb06e480e5dbe643cc281a5c12fc98117d01caefbd4a1a0ba57009eeb0fa7f3
Apache-2.0
[ "LICENSE" ]
0
2.4
dailybot-cli
0.2.0
Dailybot CLI - The command-line bridge between humans and agents. Progress reports, observability, and workflow automation for modern teams across Slack, Google Chat, Discord, MS Teams, and more.
# Dailybot CLI The command-line bridge between **humans** and **agents**. [Dailybot](https://www.dailybot.com) connects your team — whether they work in Slack, Google Chat, Discord, Microsoft Teams, or the web — with AI agents and automated workflows. The CLI brings that power to your terminal: progress reports, observability, health checks, messaging, and workflow automation for modern teams. ## Installation ```bash pip install dailybot-cli ``` Or via the install script: ```bash curl -sSL https://cli.dailybot.com/install.sh | bash ``` Requires Python 3.9+. ## For humans Authenticate once with your Dailybot email, then submit updates and check pending check-ins right from your terminal. ```bash # Log in (one-time setup, email OTP) dailybot login # See what check-ins are waiting for you dailybot status # Submit a free-text update dailybot update "Finished the auth module, starting on tests." # Or use structured fields dailybot update --done "Auth module" --doing "Tests" --blocked "None" ``` Run `dailybot` with no arguments to enter **interactive mode** — if you're not logged in yet, it will walk you through authentication first, then let you submit updates step by step. ## For agents Any software agent — AI coding assistants, CI jobs, deploy scripts, bots — can report activity through the CLI. This lets teams get visibility into what automated processes are doing, alongside human updates. Dailybot interconnects agents and humans with work analysis, progress reports, observability, and automations. Authenticate with any of these methods (checked in this order): ```bash # Option 1: Environment variable (CI pipelines, one-off scripts) export DAILYBOT_API_KEY=your-key # Option 2: Store the key on disk (recommended for dev machines) dailybot config key=your-key # Option 3: Use your login session (no API key needed) dailybot login ``` Then run agent commands: ```bash # Report a deployment dailybot agent update "Deployed v2.1 to staging" # Name the agent so the team knows who's reporting dailybot agent update "Built feature X" --name "Claude Code" # Include structured data dailybot agent update "Tests passed" --name "CI Bot" --json-data '{"suite": "integration", "passed": 42}' # Mark a report as a milestone dailybot agent update "Shipped v3.0" --milestone --name "Claude Code" # Add co-authors (repeatable flag or comma-separated) dailybot agent update "Paired on auth refactor" --co-authors alice@co.com --co-authors bob@co.com dailybot agent update "Paired on auth refactor" --co-authors "alice@co.com,bob@co.com" # Combine milestone and co-authors dailybot agent update "Launched new dashboard" --milestone --co-authors alice@co.com --name "Claude Code" # Report agent health dailybot agent health --ok --message "All systems go" --name "Claude Code" dailybot agent health --fail --message "DB unreachable" --name "CI Bot" # Check agent health status dailybot agent health --status --name "Claude Code" # Register a webhook to receive messages dailybot agent webhook register --url https://my-server.com/hook --secret my-token --name "Claude Code" # Unregister a webhook dailybot agent webhook unregister --name "Claude Code" # Send a message to an agent dailybot agent message send --to "Claude Code" --content "Review PR #42" dailybot agent message send --to "Claude Code" --content "Do X" --type command # List messages for an agent dailybot agent message list --name "Claude Code" dailybot agent message list --name "Claude Code" --pending ``` ## Commands | Command | Description | |---|---| | `dailybot login` | Authenticate with email OTP | | `dailybot logout` | Log out and revoke token | | `dailybot status` | Show pending check-ins for today | | `dailybot update` | Submit a check-in update (free-text or structured) | | `dailybot config` | Get, set, or remove a stored setting (e.g. API key) | | `dailybot agent update` | Submit an agent activity report (API key or login) | | `dailybot agent health` | Report or query agent health status (API key or login) | | `dailybot agent webhook register` | Register a webhook for the agent (API key or login) | | `dailybot agent webhook unregister` | Unregister the agent's webhook (API key or login) | | `dailybot agent message send` | Send a message to an agent (API key or login) | | `dailybot agent message list` | List messages for an agent (API key or login) | ### `dailybot agent update` ``` Usage: dailybot agent update [OPTIONS] CONTENT Submit an agent activity report. Options: -n, --name TEXT Agent worker name. -j, --json-data TEXT Structured JSON data to include. -m, --milestone Mark as a milestone accomplishment. -c, --co-authors TEXT Co-author email or UUID (repeatable, or comma-separated). --help Show this message and exit. ``` Run `dailybot --help` or `dailybot <command> --help` for full details on any command. ## Development ```bash pip install -e ".[dev]" pytest ``` ## License [MIT](LICENSE)
text/markdown
null
Dailybot <support@dailybot.com>
null
null
MIT
null
[ "Development Status :: 3 - Alpha", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Topic :: Software Development :: Build Tools" ]
[]
null
null
>=3.9
[]
[]
[]
[ "click>=8.1.0", "httpx>=0.25.0", "questionary>=2.0.0", "rich>=13.0.0" ]
[]
[]
[]
[ "Homepage, https://www.dailybot.com", "Documentation, https://docs.dailybot.com" ]
twine/6.2.0 CPython/3.13.7
2026-02-21T00:33:18.385524
dailybot_cli-0.2.0.tar.gz
17,018
12/2e/a6028f8417cb659f1f2e173cd9da2d7e7be3f1f445dda820a838c591b712/dailybot_cli-0.2.0.tar.gz
source
sdist
null
false
b916d8ac7c32a30dc25a7900c8015f1c
fdd91a6a90b272261a4c7a92de80226436cfc043a86c02b798e0476c9770f0de
122ea6028f8417cb659f1f2e173cd9da2d7e7be3f1f445dda820a838c591b712
null
[ "LICENSE" ]
157
2.4
proprioceptive-cradle
0.4.0
Behavioral adapter generation for language models via fiber-bundle probes
# proprioceptive-cradle Behavioral adapter generation for language models via fiber-bundle probes. ## What It Does The Cradle trains lightweight probes on your model's hidden states to measure 9 behavioral dimensions, then generates a LoRA adapter that shifts behavior toward a target profile. **All GPU computation runs locally. No model data leaves your machine.** ## Quick Start ```bash pip install proprioceptive-cradle ``` ### Activate Your License ```bash cradle activate <your-license-key> ``` Get a key at [proprioceptive.ai](https://proprioceptive.ai) ### Behavioral Scan ```bash cradle scan --model mistral ``` ### Generate Adapter ```bash cradle generate --model mistral --preset professional ``` ### Python API ```python from cradle import scan, generate report = scan("mistral") adapter_path = generate("mistral", preset="professional") ``` ## Validated Models | Model | Architecture | Params | |-------|-------------|--------| | Falcon-Mamba 7B | Mamba SSM | 7B | | LLaMA 3.1 8B | Transformer | 8B | | Mistral 7B | Transformer | 7B | | Qwen 2.5 3B | Transformer | 3B | | Qwen 2.5 7B | Transformer | 7B | | Qwen 2.5 32B | Transformer | 32B | ## Presets - **professional**: High depth/specificity, low sycophancy/hedging - **creative**: Relaxed calibration, higher verbosity tolerance - **cautious**: Maximum calibration, moderate hedging allowed - **direct**: Maximum specificity/focus, minimal everything else ## Requirements - Python 3.9+ - CUDA GPU with 8GB+ VRAM (16GB+ recommended) - PyTorch 2.0+ - Valid license key ## Author Logan Matthew Napolitano — Proprioceptive AI
text/markdown
null
Logan Matthew Napolitano <logan@proprioceptive.ai>
null
null
Proprietary
adapter, behavioral-probes, fiber-bundle, llm, lora
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Science/Research", "Programming Language :: Python :: 3", "Topic :: Scientific/Engineering :: Artificial Intelligence" ]
[]
null
null
>=3.9
[]
[]
[]
[ "bitsandbytes>=0.41", "click>=8.0", "datasets>=2.0", "peft>=0.7", "rich>=13.0", "torch>=2.0", "transformers>=4.36" ]
[]
[]
[]
[ "Homepage, https://proprioceptive.ai", "Repository, https://github.com/ProprioceptiveAI/cradle" ]
twine/6.2.0 CPython/3.11.14
2026-02-21T00:33:08.287374
proprioceptive_cradle-0.4.0.tar.gz
37,514
ff/72/8de978953751329cf0ffd629b198d0eac2efd77e8e02deee6a0311f7163b/proprioceptive_cradle-0.4.0.tar.gz
source
sdist
null
false
7cd95ad956df8fc1e4ac22d2ed1ab013
1b9a5c11605a0feccd1c4ea269655d9ac22049b88297546791ca906d350a597e
ff728de978953751329cf0ffd629b198d0eac2efd77e8e02deee6a0311f7163b
null
[]
144
2.4
Geode-Common
33.19.0
Common module for licensed Geode-solutions modules
<h1 align="center">Geode-Common<sup><i>by Geode-solutions</i></sup></h1> <h3 align="center">Common module for proprietary Geode-solutions modules</h3> <p align="center"> <img src="https://github.com/Geode-solutions/Geode-Common_private/workflows/CI/badge.svg" alt="Build Status"> <img src="https://github.com/Geode-solutions/Geode-Common_private/workflows/CD/badge.svg" alt="Deploy Status"> <img src="https://img.shields.io/github/release/Geode-solutions/Geode-Common.svg" alt="Version"> <img src="https://img.shields.io/pypi/v/geode-common" alt="PyPI" > </p> <p align="center"> <img src="https://img.shields.io/static/v1?label=Windows&logo=windows&logoColor=white&message=support&color=success" alt="Windows support"> <img src="https://img.shields.io/static/v1?label=Ubuntu&logo=Ubuntu&logoColor=white&message=support&color=success" alt="Ubuntu support"> <img src="https://img.shields.io/static/v1?label=Red%20Hat&logo=Red-Hat&logoColor=white&message=support&color=success" alt="Red Hat support"> </p> <p align="center"> <img src="https://img.shields.io/badge/C%2B%2B-17-blue.svg" alt="Language"> <img src="https://img.shields.io/badge/%20%20%F0%9F%93%A6%F0%9F%9A%80-semantic--release-e10079.svg" alt="Semantic-release"> <a href="https://geode-solutions.com/#slack"> <img src="https://opengeode-slack-invite.herokuapp.com/badge.svg" alt="Slack invite"> </a> </p>
text/markdown
null
Geode-solutions <contact@geode-solutions.com>
null
null
Proprietary
null
[]
[]
null
null
<3.13,>=3.9
[]
[]
[]
[ "opengeode-core==15.*,>=15.31.5" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.9.13
2026-02-21T00:32:16.020354
geode_common-33.19.0-cp39-cp39-win_amd64.whl
6,573,714
a7/48/bbc2a33c2533e5f41528c82cfe616bc1e96e41b3d655013a2f4ca12f7127/geode_common-33.19.0-cp39-cp39-win_amd64.whl
cp39
bdist_wheel
null
false
8e0c4b20f84fa009a67caaa0f3748762
f3caefd15cd109749817c68e0ef46bbddcbc963d1e322a9400bf12a78a8f4ac3
a748bbc2a33c2533e5f41528c82cfe616bc1e96e41b3d655013a2f4ca12f7127
null
[]
0
2.4
fbtk
0.9.7
High-performance molecular simulation toolkit with Python bindings
# FBTK: Forblaze Toolkit ![PyPI - Version](https://img.shields.io/pypi/v/fbtk) [![Documentation](https://img.shields.io/badge/docs-latest-blue)](https://fbtk.forblaze-works.com/) High-performance molecular analysis and building tools powered by Rust. Designed as a "Transparent Accelerator" for Python (ASE/RDKit) workflows with a smart, object-oriented interface. ## Features - **🚀 High Performance**: Core logic written in Rust with parallel processing (Rayon). - **🏗️ Intelligent Builder**: - **Optimized Initial Packing**: Grid-based placement with uniform density. - **Polymer Synthesis**: Automatic chain generation with leaving atom support. - **Tacticity Control**: Supports **Isotactic**, **Syndiotactic**, and **Atactic** arrangements via center-of-mass analysis and mirroring. - **Built-in 3D Generation**: 3D coordinate generation from SMILES handled by internal VSEPR + UFF engine. - **Fast Structural Relaxation**: O(N) Cell-list optimization with the FIRE algorithm. - **🔍 Advanced Analysis**: - Parallel RDF, MSD, COM (Center of Mass), Angles, Dihedrals. - O(N) Neighbor List search. - **📏 Robust Physics**: Correct handling of PBC, Triclinic cells, and Minimum Image Convention (MIC). ## Installation ### Python Library (Recommended) ```bash pip install fbtk ``` **Requirements**: Python 3.8+ (Zero external dependencies). ### Standalone CLI (No Python Required) For non-Python environments, pre-compiled standalone binaries for Linux, Windows, and macOS are available on the [GitHub Releases](https://github.com/ForblazeProject/fbtk/releases) page. - Download the archive for your platform (e.g., `fbtk-cli-v0.9.6-linux-x86_64.tar.gz`). - **Requirements**: None. These are self-contained binaries. ## Key Features - **🚀 High Performance**: Core logic written in Rust with parallel processing (Rayon). - **🏗️ Intelligent Builder**: - **Optimized Initial Packing**: Grid-based placement with uniform density. - **Polymer Synthesis**: Automatic chain generation with terminal capping (hydrogen addition). - **Tacticity Control**: Supports **Isotactic**, **Syndiotactic**, and **Atactic** arrangements. - **Charge Assignment**: Automatic partial charge calculation via Gasteiger method. - **Built-in 3D Generation**: 3D coordinate generation from SMILES handled by internal VSEPR + UFF engine. - **Fast Structural Relaxation**: O(N) Cell-list optimization with the FIRE algorithm. - **🔍 Advanced Analysis**: - Parallel RDF, MSD, COM (Center of Mass), Angles, Dihedrals. - **📏 Robust Physics**: Correct handling of PBC, Triclinic cells, and Minimum Image Convention (MIC). - **📦 Zero Dependency**: No RDKit, no NumPy, no SciPy required for the core engine. Perfect for clean deployment. ### 1. System Building Build and relax a complex molecular system with just a few lines of code. ```python import fbtk # 1. Setup Builder builder = fbtk.Builder(density=0.8) builder.add_molecule_smiles("ethanol", count=50, smiles="CCO") # 2. Build and Relax system = builder.build() system.relax(steps=500) # 3. Export to ASE atoms = system.to_ase() atoms.write("system.xyz") ``` ### 2. RDF Analysis Fast analysis of large trajectories using smart selection queries. ```python from ase.io import read import fbtk # Load trajectory (ASE list of Atoms) traj = read('simulation.lammpstrj', index=':') # Compute RDF using a simple query string r, g_r = fbtk.compute_rdf(traj, query="O-H", r_max=10.0) ``` --- ### 3. Command Line Interface (CLI) FBTK provides standalone CLI tools for batch processing. #### fbtk-build: Build from Recipe ```bash # Run building and relaxation from a YAML recipe fbtk-build --recipe recipe.yaml --relax --output system.mol2 ``` Example `recipe.yaml`: ```yaml system: density: 0.8 cell_shape: [20.0, 20.0, 20.0] components: - name: "ethanol" role: "molecule" input: smiles: "CCO" count: 50 ``` #### fbtk-analyze: Analyze Trajectory ```bash # Compute RDF for a LAMMPS trajectory fbtk-analyze rdf --input traj.lammpstrj --query "type 1 with type 2" ``` ## Selection Query Syntax FBTK supports intuitive strings to select atoms for analysis: - **Element**: `"O"`, `"H"`, `"element C"` - **Pairs (RDF)**: `"O-H"`, `"C - C"` - **Index Range**: `"index 0:100"` (start:end) - **Residue**: `"resname STY"` --- ### Author **Forblaze Project** Website: [https://forblaze-works.com/en/](https://forblaze-works.com/en/)
text/markdown; charset=UTF-8; variant=GFM
Forblaze Project Contributors
null
null
null
MIT OR Apache-2.0
molecular-dynamics, simulation, builder, cheminformatics
[ "Programming Language :: Rust", "Programming Language :: Python :: 3", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Topic :: Scientific/Engineering :: Chemistry", "License :: OSI Approved :: MIT License", "License :: OSI Approved :: Apache Software License" ]
[]
null
null
>=3.8
[]
[]
[]
[]
[]
[]
[]
[]
uv/0.10.4 {"installer":{"name":"uv","version":"0.10.4","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
2026-02-21T00:31:42.875592
fbtk-0.9.7-pp310-pypy310_pp73-manylinux_2_17_armv7l.manylinux2014_armv7l.whl
1,125,105
74/9d/05b6b24c61ca5c0c6b4451da69fcb8abe4a48334d1fdaaffc548c30201ff/fbtk-0.9.7-pp310-pypy310_pp73-manylinux_2_17_armv7l.manylinux2014_armv7l.whl
pp310
bdist_wheel
null
false
d2f0371cd5dcbbc3b7e7794806148971
f54d8dce0e3f0f16967253dca1cdc10836c59a2edb0caa1e138426af37b10987
749d05b6b24c61ca5c0c6b4451da69fcb8abe4a48334d1fdaaffc548c30201ff
null
[ "LICENSE-APACHE", "LICENSE-MIT" ]
470
2.4
pygpubench
0.0.2
GPU kernel benchmarking utilities
# PyGPUBench Utilities for benchmarking low-latency CUDA kernels in an _adversarial_ setting. Contrary to many existing benchmarking tools, which generally assume a cooperative kernel that can be tested and benchmarked independently, this library tries to defend against kernels that try to exploit benchmarking flaws to receive higher scores. ## Usage To benchmark a kernel, two ingredients are needed: 1. A function that _generates_ the kernel. This function takes no arguments and returns a callable. It is important that untrusted code, e.g., the user-supplied python module, is only imported inside this function. 2. A function that generates test/benchmark inputs. This function takes a tuple of configuration parameters, as well as an integer to seed the rng, as arguments. It returns two tuples: The first contains the inputs for the kernel and will be used to call the kernel function, and the second contains the expected output and the required absolute and relative tolerance. ```python import torch import pygpubench def generate_input(*args): ... def reference_kernel(args): ... def generate_test_case(args, seed): x, y = generate_input(*args, seed) expected = torch.empty_like(y) reference_kernel((expected, x)) return (y, x), (expected, 1e-6, 1e-6) def kernel_generator(): import submission return submission.kernel res = pygpubench.do_bench_isolated(kernel_generator, generate_test_case, (1024,), 100, 5, discard=True) print("❌" if res.errors else "✅", pygpubench.basic_stats(res.time_us)) ``` For the full example see [grayscale.py](test/grayscale.py) ## Implementation Unfortunately, any benchmarking tool written in python is inherently vulnerable to monkeypatching and `inpect`-based manipulation of its variables by its callees. Therefore, PyGPUBench implements its main benchmarking logic in a compiled C++ extension. While this still leaves vulnerabilities - the code is running in the same address space, after all – it makes attacks require much more sophistication. Running in a separate process fundamentally clashes with the desire to benchmark very short kernels; cuda events must be recorded in the same process as the kernel. Fortunately, we can assume that a reward-hacking LLM is still rather unlikely to produce a compiled extension that runs sophisticated low-level exploits. Note that, as soon as any user code is executed, the entire python runtime becomes untrustworthy. Consequently, benchmark results are not returned to python, but instead written to a file. The name of this file is passed as an argument to the benchmarking function, and the file is unlinked before the user code is called, making it impossible to reopen this file. The `do_bench_isolated` function is designed to streamline this process: It automates creating the temporary file, spawning a new python process to handle benchmarking and reading the results back into python (the original, untainted process). Thus, the library provides two main interfaces to benchmarking: `do_bench_impl` runs benchmarking directly in the current process, `do_bench_isolated` runs it in a separate process and automaticallly handles I/O through a temporary file. Additional measures to mitigate benchmark cheating are that benchmark inputs are generated before any benchmark is run, but then moved to a GPU memory location unknown to `torch` (allocated directly with cudaMalloc in C++). Only before the actual kernel is launched do we copy the inputs back to their original locations. Problematically, this would put the inputs into L2 cache, which we want to avoid. This means that between the copy and the kernel launch, there has to be another kernel that clears the L2 cache, opening a window of opportunity for cheating. To minimize the duration of vulnerability, we put a small fraction of random canaries into the input data, that is, a subset of memory location contains wrong data. Only after L2 clearing do we fix up these values; this pulls them into L2 cache, but since they make up less than 1% of the total data, we consider this an acceptable tradeoff. Similarly, after the kernel is finished, we directly launch the testing kernel with a programmatically-dependent launch, again to minimize the window of opportunity for cheating by writing results from a different stream. This could have a small effect on performance, as during the tail of the user kernel blocks of the test kernel are already put on the SMs and generate memory traffic. In the checking kernel, the order in which blocks are checked is randomized, so that it is not a viable strategy to only write the later blocks of the result from an unsynchronized stream.
text/markdown
null
Erik Schultheis <erik.schultheis@ist.ac.at>
null
null
null
null
[ "Development Status :: 3 - Alpha", "Environment :: GPU :: NVIDIA CUDA :: 13", "Operating System :: POSIX :: Linux", "Programming Language :: C++", "Programming Language :: Python :: 3" ]
[]
null
null
>=3.12
[]
[]
[]
[]
[]
[]
[]
[ "Repository, https://github.com/ngc92/pygpubench", "Issues, https://github.com/ngc92/pygpubench/issues" ]
twine/6.2.0 CPython/3.13.3
2026-02-21T00:31:29.468043
pygpubench-0.0.2-cp312-abi3-manylinux_2_35_x86_64.whl
1,056,219
6e/da/733c4c516c6a1d60fd530c59dab2ff7f6d3c64d325e26ea0da2c7203ed85/pygpubench-0.0.2-cp312-abi3-manylinux_2_35_x86_64.whl
cp312
bdist_wheel
null
false
9511570f0f56ee94a2fc01d55279548c
0d38c3e84abec932a51b5f3dcc736b67446cc0e558d1f66ccd283416bdb4707a
6eda733c4c516c6a1d60fd530c59dab2ff7f6d3c64d325e26ea0da2c7203ed85
Apache-2.0
[]
72
2.4
django-spire
0.28.0
A project for Django Spire
# Django Spire ![Build](https://img.shields.io/github/actions/workflow/status/stratusadv/django-spire/ci.yml) ![Python Versions](https://img.shields.io/pypi/pyversions/django-spire) ![PyPI Version](https://img.shields.io/pypi/v/django-spire) ![Downloads](https://img.shields.io/pypi/dm/django-spire) ## Purpose Make Django application development modular, extensible, and configurable. ## Getting Started Please refer to our [website](https://django-spire.stratusadv.com) for detailed setup, module configuration, and usage examples. <p align="center"> <a href="https://django-spire.stratusadv.com"> <img alt="Dandy Logo" src="https://django-spire.stratusadv.com/static/img/django_spire_logo_512.png"/> </a> </p>
text/markdown
null
Brayden Carlson <braydenc@stratusadv.com>, Nathan Johnson <nathanj@stratusadv.com>
null
null
Copyright (c) 2025 Stratus Advanced Technologies and Contributors. Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
portal, cmms, spire, django, backend, frontend, javascript, active server pages
[ "Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Django", "Framework :: Django :: 3.2", "Framework :: Django :: 4.0", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: JavaScript", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Topic :: Internet :: WWW/HTTP", "Topic :: Internet :: WWW/HTTP :: Dynamic Content", "Topic :: Software Development :: Libraries :: Python Modules" ]
[]
null
null
>=3.11
[]
[]
[]
[ "beautifulsoup4>=4.14.2", "boto3>=1.34.0", "botocore>=1.34.0", "crispy-bootstrap5==2024.10", "dandy==2.0.0", "django>=5.1.8", "django-crispy-forms==2.3", "django-glue>=0.8.12", "django-sendgrid-v5", "django-storages==1.14.5", "docutils", "faker", "marko==2.1.4", "markitdown[docx,pdf]==0.1.2", "psycopg2", "pydantic", "Pygments", "Pillow==11.2.1", "python-dateutil", "robit>=0.4.8", "typing_extensions", "twilio", "build; extra == \"development\"", "django-browser-reload; extra == \"development\"", "django-debug-toolbar; extra == \"development\"", "django-watchfiles; extra == \"development\"", "pip; extra == \"development\"", "playwright; extra == \"development\"", "pyinstrument; extra == \"development\"", "pyrefly; extra == \"development\"", "pytest; extra == \"development\"", "pytest-django; extra == \"development\"", "pytest-playwright; extra == \"development\"", "setuptools; extra == \"development\"", "twine; extra == \"development\"", "mkdocs; extra == \"documentation\"", "markdown-exec[ansi]; extra == \"documentation\"", "mkdocs-include-markdown-plugin; extra == \"documentation\"", "mkdocs-material; extra == \"documentation\"", "mkdocs-table-reader-plugin; extra == \"documentation\"", "mkdocstrings-python; extra == \"documentation\"", "mkdocs-gen-files; extra == \"documentation\"", "mkdocs-literate-nav; extra == \"documentation\"", "mkdocs-section-index; extra == \"documentation\"", "openpyxl; extra == \"documentation\"" ]
[]
[]
[]
[]
uv/0.10.4 {"installer":{"name":"uv","version":"0.10.4","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
2026-02-21T00:31:27.080297
django_spire-0.28.0-py3-none-any.whl
2,532,743
d0/23/05c1db1129a0a4b04a20ceee95dcf202786adca4ff35d988b0563cca306e/django_spire-0.28.0-py3-none-any.whl
py3
bdist_wheel
null
false
fbbf68442a77c1cd57410ac0447ffb54
9fa3e71f24de8fc9883cd8906657ccf6f721d6e1fbbeffd3a7790fb037dfa959
d02305c1db1129a0a4b04a20ceee95dcf202786adca4ff35d988b0563cca306e
null
[ "LICENSE.md" ]
173
2.4
ddapm-test-agent
1.42.0
Test agent for Datadog APM client libraries
# Datadog APM test agent [![GitHub Workflow Status (with branch)](https://img.shields.io/github/actions/workflow/status/DataDog/dd-apm-test-agent/main.yml?style=flat-square)](https://github.com/DataDog/dd-apm-test-agent/actions?query=workflow%3ACI+branch%3Amaster) [![PyPI](https://img.shields.io/pypi/v/ddapm-test-agent?style=flat-square)](https://pypi.org/project/ddapm-test-agent/) <img align="right" src="https://user-images.githubusercontent.com/6321485/136316621-b4af42b6-4d1f-4482-a45b-bdee47e94bb8.jpeg" alt="bits agent" width="200px"/> The APM test agent is an application which emulates the APM endpoints of the [Datadog agent](https://github.com/DataDog/datadog-agent/) which can be used for testing Datadog APM client libraries. See the [Features](#features) section for the complete list of functionalities provided. See the [HTTP API](#http-api) section for the endpoints available. See the [Development](#development) section for how to get the test agent running locally to add additional checks or fix bugs. ## Installation The test agent can be installed from PyPI: pip install ddapm-test-agent # HTTP on port 8126, OTLP HTTP on port 4318, OTLP GRPC on port 4317, with the web-ui enabled ddapm-test-agent --port=8126 --otlp-http-port=4318 --otlp-grpc-port=4317 --web-ui-port=8080 or from Docker: # Run the test agent and mount the snapshot directory docker run --rm\ -p 8126:8126\ -p 4318:4318\ -p 4317:4317\ -e SNAPSHOT_CI=0\ -v $PWD/tests/snapshots:/snapshots\ ghcr.io/datadog/dd-apm-test-agent/ddapm-test-agent:latest or from source: pip install git+https://github.com/Datadog/dd-apm-test-agent or a specific branch: pip install git+https://github.com/Datadog/dd-apm-test-agent@{branch} ## Features ### Trace invariant checks Many checks are provided by the test agent which will verify trace data. All checks are enabled by default and can be manually disabled. See the [configuration](#configuration) section for the options. | Check description | Check name | | ------------- | ------------- | | Trace count header matches number of traces | `trace_count_header` | | Client library version header included in request | `meta_tracer_version_header` | | Trace content length header matches payload size | `trace_content_length` | ### Returning data All data that is submitted to the test agent can be retrieved. - Traces can be returned via the `/test/traces` endpoint documented [below](#api). ### Helpful logging The `INFO` log level of the test agent outputs useful information about the requests the test agent receives. For traces this includes a visual representation of the traces. ``` INFO:ddapm_test_agent.agent:received trace payload with 1 trace chunk INFO:ddapm_test_agent.agent:Chunk 0 [parent] ├─ [child1] ├─ [child2] └─ [child3] INFO:ddapm_test_agent.agent:end of payload ---------------------------------------- ``` ### Proxy The test agent provides proxying to the Datadog agent. This is enabled by passing the agent url to the test agent either via the `--agent-url` command-line argument or by the `DD_TRACE_AGENT_URL` or `DD_AGENT_URL` environment variables. When proxying is enabled, the response from the Datadog agent will be returned instead of one from the test agent. At the trace-level, proxying can also be disabled by including the `X-Datadog-Agent-Proxy-Disabled` header with a value of `true`. This will disable proxying after a trace is handled, regardless of whether an agent URL is set. ### LLM Observability Proxy The test agent provides a mechanism to dual ship LLM Observability events to Datadog, regardless of whether the Datadog agent is running. If using the Datadog agent, set the `DD_AGENT_URL` environment variable or `--agent-url` command-line argument to the URL of the Datadog agent (see [Proxy](#proxy) for more details). If not running a Datadog agent, set the `DD_SITE` environment variable or `--dd-site` command-line argument to the site of the Datadog instance to forward events to. Additionally, set the `DD_API_KEY` environment variable or `--dd-api-key` command-line argument to the API key to use for the Datadog instance. To disable LLM Observability event forwarding, set the `DISABLE_LLMOBS_DATA_FORWARDING` environment variable or `--disable-llmobs-data-forwarding` command-line argument to `true`. ### Claude Code Hooks The test agent can receive [Claude Code hook](https://docs.claude.com/en/docs/claude-code/hooks) events and assemble them into LLM Observability traces. See [`.claude/setup-hooks.md`](.claude/setup-hooks.md) for full setup instructions (also usable as a Claude Code prompt to automate the setup). ### Snapshot testing The test agent provides a form of [characterization testing](https://en.wikipedia.org/wiki/Characterization_test) which we refer to as snapshotting. This allows library maintainers to ensure that traces don't change unexpectedly when making unrelated changes. This can be used to write integration tests by having test cases use the tracer to emit traces which are collected by the test agent and compared against reference traces stored previously. To do snapshot testing with the test agent: 1. Ensure traces are associated with a session token (typically the name of the test case) by either: - Calling the `/test/session/start` with the token endpoint before emitting the traces; or - Attaching an additional query string parameter or header specifying the session token on `/vX.Y/trace` requests (see below for the API specifics). (Required for concurrent test running) 2. Emit traces (run the integration test). 3. Signal the end of the session and perform the snapshot comparison by calling the `/tests/session/snapshot` endpoint with the session token. The endpoint will return a `400` response code if the snapshot failed along with a plain-text trace of the error which can be forwarded to the test framework to help triage the issue. #### Snapshot output The traces are normalized and output in JSON to a file. The following transformations are made to the input: - Trace ids are overwritten to match the order in which the traces were received. - Span ids are overwritten to be the DFS order of the spans in the trace tree. - Parent ids are overwritten using the normalized span ids. However, if the parent is not a span in the trace, the parent id is not overwritten. This is necessary for handling distributed traces where all spans are not sent to the same agent. - Span attributes are ordered to be more human-readable, with the important attributes being listed first. - Span attributes are otherwise ordered alphanumerically. - The span meta and metrics maps if empty are excluded. #### Web UI The test agent includes an optional and **experimental** Web UI that provides a dashboard for inspecting agent configuration, viewing received requests, exploring traces and snapshots, and managing tracer-flare and remote configuration. The UI can be enabled with the `--web-ui-port PORT` command-line argument or by setting the `WEB_UI_PORT` environment variable. Once enabled, the Web UI can be accessed at `http://localhost:PORT` (default port is `8080`). There is also a maximum number of requests to store in memory to display in the UI, which can be configured with the `--max-requests` command-line argument or by setting the `MAX_REQUESTS` environment variable (default is `200` requests). ### Recording 3rd party API requests The test agent can be configured to proxy requests to select provider API endpoints, capturing real requests to the server and recording them to play back for future use. Currently, only OpenAI, Azure OpenAI, and DeepSeek are supported. These cassettes are recorded by default in the `vcr-cassettes` directory. However, this can be changed with the `--vcr-cassettes-directory` command-line option, or `VCR_CASSETTES_DIRECTORY` environment variable. The cassettes are matched based on the path, method, and body of the request. To mount a cassette directory when running the test agent in a Docker container, run the container with docker run --rm\ -p 8126:8126\ -v $PWD/vcr-cassettes:/vcr-cassettes ghcr.io/datadog/dd-apm-test-agent/ddapm-test-agent:latest Optionally specifying whatever mounted path is used for the cassettes directory. The test agent comes with a default set of cassettes for OpenAI, Azure OpenAI, DeepSeek, Anthropic, Google GenAI, and AWS Bedrock Runtime. #### Custom 3rd Party Providers The test agent can be configured to also register custom 3rd party providers. This is done by setting the `VCR_PROVIDER_MAP` environment variable or the `--vcr-provider-map` command-line option to a comma-separated list of provider names and their corresponding base URLs. ```shell VCR_PROVIDER_MAP="provider1=http://provider1.com/,provider2=http://provider2.com/" ``` or ```shell --vcr-provider-map="provider1=http://provider1.com/,provider2=http://provider2.com/" ``` The provider names are used to match the provider name in the request path, and the base URLs are used to proxy the request to the corresponding provider API endpoint. With this configuration set, you can make the following request to the test agent without error: ```shell curl -X POST 'http://127.0.0.1:9126/vcr/provider1/some/path' ``` #### Ignoring Headers in Recorded Cassettes To ignore headers in recorded cassettes, you can use the `--vcr-ignore-headers` flag or `VCR_IGNORE_HEADERS` environment variable. The list should take the form of `header1,header2,header3`, and will be omitted from the recorded cassettes. #### AWS Services AWS service proxying, specifically recording cassettes for the first time, requires a `AWS_SECRET_ACCESS_KEY` environment variable to be set for the container running the test agent. This is used to recalculate the AWS signature for the request, as the one generated client-side likely used `{test-agent-host}:{test-agent-port}/vcr/{aws-service}` as the host, and the signature will mismatch that on the actual AWS service. Additionally, the `AWS_REGION` environment variable can be set, defaulting to `us-east-1`. To add a new AWS service to proxy, add an entry in the `PROVIDER_BASE_URLS` for its provider url, and an entry in the `AWS_SERVICES` dictionary for the service name, since they are not always a one-to-one mapping with the implied provider url (e.g, `https://bedrock-runtime.{AWS_REGION}.amazonaws.com` is the provider url, but the service name is `bedrock`, as `bedrock` also has multiple sub services, like `converse`). #### Usage in clients To use this feature in your client, you can use the `/vcr/{provider}` endpoint to proxy requests to the provider API. ```python from openai import OpenAI client = OpenAI(base_url="http://127.0.0.1:9126/vcr/openai") ``` #### Recording test names as part of VCR cassette names The test agent has two endpoints to configure a context around which any VCR cassettes recorded have a suffix of a given test name. To use this, you can hit the - `/vcr/test/start` with a `test_name` body field to set the test name - `/vcr/test/stop` to clear the test name This is useful for recording cassettes for a specific test case to easily associate cassettes with that test. Usage example: ```python @pytest.mark.fixture def with_vcr_test_name(request): with requests.post("http://127.0.0.1:9126/vcr/test/start", json={"test_name": request.node.name}): yield requests.post("http://127.0.0.1:9126/vcr/test/stop") @pytest.mark.fixture def openai_with_custom_url(with_vcr_test_name): client = OpenAI(base_url="http://127.0.0.1:9126/vcr/openai") yield client def test_openai_with_custom_url(openai_with_custom_url): """This test will generate/use a cassette name something similar to `openai_chat_completions_post_abcd1234_test_openai_with_custom_url`""" ... ``` #### Adding new providers To add a new provider, add a supported provider in the `PROVIDER_BASE_URLS` dictionary in `ddapm_test_agent/vcr_proxy.py`, and change your tests or use case to use the new provider in the base url: ```python base_url = "http://127.0.0.1:9126/vcr/{new_provider}" ``` And pass in a valid API key (if needed) in the way that provider expects. To redact api keys, modify the `filter_headers` list in the `get_vcr` function in `ddapm_test_agent/vcr_proxy.py`. This can be confirmed by viewing cassettes in the `vcr-cassettes` directory (or the otherwise specified directory), and verifying that any new cassettes do not contain the api key. #### Running in CI To have the vcr proxy throw a 404 error if a cassette is not found in CI mode to ensure that all cassettes are generated locally and committed, set the `VCR_CI_MODE` environment variable or the `--vcr-ci-mode` flag in the cli tool to `true` (this value defaults to `false`). ## Configuration The test agent can be configured via command-line options or via environment variables. ### Command line #### ddapm-test-agent `ddapm-test-agent` is command used to run a test agent. Please refer to `ddapm-test-agent --help` for more information. #### ddapm-test-agent-fmt `ddapm-test-agent-fmt` is a command line tool to format or lint snapshot json files. ``` bash # Format all snapshot json files ddapm-test-agent-fmt path/to/snapshots # Lint snapshot json files ddapm-test-agent-fmt --check path/to/snapshots ``` Please refer to `ddapm-test-agent-fmt --help` for more information. ### Environment Variables - `PORT` [`8126`]: Port to listen on. - `ENABLED_CHECKS` [`""`]: Comma-separated values of checks to enable. Valid values can be found in [trace invariant checks](#trace-invariant-checks) - `LOG_LEVEL` [`"INFO"`]: Log level to use. DEBUG, INFO, WARNING, ERROR, CRITICAL. - `LOG_SPAN_FMT` [`"[{name}]"`]: Format string to use when outputting spans in logs. - `SNAPSHOT_DIR` [`"./snapshots"`]: Directory in which snapshots will be stored. Can be overridden by providing the `dir` query parameter on `/snapshot`. - `SNAPSHOT_CI` [`0`]: Toggles CI mode for the snapshot tests. Set to `1` to enable. CI mode does the following: - When snapshots are unexpectedly _generated_ from a test case a failure will be raised. - `SNAPSHOT_IGNORED_ATTRS` [`"span_id,trace_id,parent_id,duration,start,metrics.system.pid,metrics.process_id,metrics.system.process_id,meta.runtime-id"`]: The attributes to ignore when comparing spans in snapshots. - `DD_AGENT_URL` [`""`]: URL to a Datadog agent. When provided requests will be proxied to the agent. - `DD_APM_RECEIVER_SOCKET` [`""`]: When provided, the test agent will listen for traces on a socket at the path provided (e.g., `/var/run/datadog/apm.socket`) - `DD_SUPPRESS_TRACE_PARSE_ERRORS` [`false`]: Set to `"true"` to disable span parse errors when decoding handled traces. When disabled, errors will not be thrown for metrics incorrectly placed within the meta field, or other type errors related to span tag formatting/types. Can also be set using the `--suppress-trace-parse-errors=true` option. - `SNAPSHOT_REMOVED_ATTRS` [`""`]: The attributes to remove from spans in snapshots. This is useful for removing attributes that are not relevant to the test case. **Note that removing `span_id` is not permitted to allow span ordering to be maintained.** - `SNAPSHOT_REGEX_PLACEHOLDERS` [`""`]: The regex expressions to replace by a placeholder. Expressed as a comma separated `key:value` list. Specifying `ba[rz]:placeholder` will change any occurrence of `bar` or `baz` to `{placeholder}`: `foobarbazqux` -> `foo{placeholder}{placeholder}qux`. This is in particular useful to strip path prefixes or other infrastructure dependent identifiers. - `DD_POOL_TRACE_CHECK_FAILURES` [`false`]: Set to `"true"` to pool Trace Check failures that occured within Test-Agent memory. These failures can be queried later using the `/test/trace_check/failures` endpoint. Can also be set using the `--pool-trace-check-failures=true` option. - `DD_DISABLE_ERROR_RESPONSES` [`false`]: Set to `"true"` to disable Test-Agent `<Response 400>` when a Trace Check fails, instead sending a valid `<Response 200>`. Recommended for use with the `DD_POOL_TRACE_CHECK_FAILURES` env variable. Can also be set using the `--disable-error-responses=true` option. ## HTTP API ### /test/traces Return traces that have been received by the agent. Traces matching specific trace ids can be requested with the options below. #### [optional] `?trace_ids=` #### [optional] `X-Datadog-Trace-Ids` Specify trace ids as comma separated values (eg. `12345,7890,2468`) ### /test/session/start Initiate a _synchronous_ session. All subsequent traces received will be associated with the required test token provided. #### [optional] `?agent_sample_rate_by_service=` Sample rates to be returned by the agent in response to trace v0.4 and v0.5 requests. Example: `"{'service:test,env:staging': 0.5, 'service:test2,env:prod': 0.2}"` (note the JSON has to be URL-encoded). #### [optional] `?test_session_token=` #### [optional] `X-Datadog-Test-Session-Token` Test session token for a test case. **Ensure this value is unique to avoid conflicts between sessions.** ### /test/session/snapshot Perform a snapshot generation or comparison on the data received during the session. Snapshots are generated when the test agent is not in CI mode and there is no snapshot file present. Otherwise a snapshot comparison will be performed. #### [optional\*] `?test_session_token=` #### [optional\*] `X-Datadog-Test-Session-Token` To run test cases in parallel this HTTP header must be specified. All test cases sharing a test token will be grouped. \* Required for concurrent tests. Either via query param or HTTP header. #### [optional] `?ignores=` Comma-separated list of keys of which to ignore values for. The default built-in ignore list is: `span_id`, `trace_id`, `parent_id`, `duration`, `start`, `metrics.system.pid`, `metrics.process_id`, `metrics.system.process_id`, `meta.runtime-id`. #### [optional] `?dir=` default: `./snapshots` (relative to where the test agent is run). Override the directory where the snapshot will be stored and retrieved from. **This directory must already exist**. This value will override the environment variable `SNAPSHOT_DIR`. Warning: it is an error to specify both `dir` and `file`. #### [optional] `?file=` #### [optional] `X-Datadog-Test-Snapshot-Filename` An absolute or relative (to the current working directory of the agent) file name where the snap will be stored and retrieved. Warning: it is an error to specify both `file` and `dir`. Note: the file extension will be appended to the filename. `_tracestats` will be appended to the filename for trace stats requests. #### [optional] `?removes=` Comma-separated list of keys that will be removed from spans in the snapshot. The default built-in remove list does not remove any keys. ### /test/session/requests Return all requests that have been received by the agent for the given session token. #### [optional] `?test_session_token=` #### [optional] `X-Datadog-Test-Session-Token` Returns the requests in the following json format: ```json [ { "headers": {}, "body": "...", "url": "http...", "method": "GET" } ] ``` `body` is a base64 encoded body of the request. ### /test/session/traces Return traces that have been received by the agent for the given session token. #### [optional] `?test_session_token=` #### [optional] `X-Datadog-Test-Session-Token` ### /test/session/stats Return stats that have been received by the agent for the given session token. #### [optional] `?test_session_token=` #### [optional] `X-Datadog-Test-Session-Token` Stats are returned as a JSON list of the stats payloads received. ### /test/session/logs Return OpenTelemetry logs that have been received by the agent for the given session token. #### [optional] `?test_session_token=` #### [optional] `X-Datadog-Test-Session-Token` Logs are returned as a JSON list of the OTLP logs payloads received. The logs are in the standard OpenTelemetry Protocol (OTLP) v1.7.0 format, decoded from protobuf into JSON. ### /test/session/metrics Return OpenTelemetry metrics that have been received by the agent for the given session token. #### [optional] `?test_session_token=` #### [optional] `X-Datadog-Test-Session-Token` Metrics are returned as a JSON list of the OTLP metrics payloads received. The metrics are in the standard OpenTelemetry Protocol (OTLP) v1.7.0 format, decoded from protobuf into JSON. ### /test/session/responses/config (POST) Create a Remote Config payload to retrieve in endpoint `/v0.7/config` #### [optional] `?test_session_token=` #### [optional] `X-Datadog-Test-Session-Token` ``` curl -X POST 'http://0.0.0.0:8126/test/session/responses/config' -d '{"roots": ["eyJ....fX0="], "targets": "ey...19", "target_files": [{"path": "datadog/2/ASM_DATA/blocked_users/config", "raw": "eyJydWxlc19kYXRhIjogW119"}], "client_configs": ["datadog/2/ASM_DATA/blocked_users/config"]}' ``` ### /test/session/responses/config/path (POST) Due to Remote Config payload being quite complicated, this endpoint works like `/test/session/responses/config (POST)` but you should send a path and a message and this endpoint builds the Remote Config payload. The keys of the JSON body are `path` and `msg` #### [optional] `?test_session_token=` #### [optional] `X-Datadog-Test-Session-Token` ``` curl -X POST 'http://0.0.0.0:8126/test/session/responses/config/path' -d '{"path": "datadog/2/ASM_DATA/blocked_users/config", "msg": {"rules_data": []}}' ``` ### /test/trace_check/failures (GET) Get Trace Check failures that occured. If a token is included, trace failures for only that session token are returned unless used in conjuction with `return_all`, which can be used to return all failures regardless of inputted token. This method returns a `<Response 200>` if no Trace Check failures are being returned and a `<Response 400>` if Trace Check failures are being returned. Trace Check failures are returned as a content type of text, with failure messages concatenated in the response body. Optionally, set the `use_json` query string parameter to `true` to return Trace Check failures as a JSON response in the following format: ``` response = { "<FAILING_CHECK_NAME>" : ["<FAILURE_MESSAGE_1>", "<FAILURE_MESSAGE_2>"] } ``` NOTE: To be used in combination with `DD_POOL_TRACE_CHECK_FAILURES`, or else failures will not be saved within Test-Agent memory and a `<Response 200>` will always be returned. #### [optional] `?test_session_token=` #### [optional] `X-Datadog-Test-Session-Token` #### [optional] `?use_json=` #### [optional] `?return_all=` ``` curl -X GET 'http://0.0.0.0:8126/test/trace_check/failures' ``` ### /test/trace_check/clear (GET) Clear Trace Check failures that occured. If a token is included, trace failures for only that session token are cleared unless used in conjuction with `clear_all`. This argument can be used to clear all failures (regardless of inputted session token). #### [optional] `?test_session_token=` #### [optional] `X-Datadog-Test-Session-Token` #### [optional] `?clear_all=` ``` curl -X GET 'http://0.0.0.0:8126/test/trace_check/clear' ``` ### /test/trace_check/summary (GET) Get Trace Check summary results. If a token is included, returns summary results only for Trace Checks run during the session. The `return_all` optional query string parameter can be used to return all trace check results (regardless of inputted session token). The method returns Trace Check results in the following JSON format: ``` summary = { "trace_content_length" : { "Passed_Checks": 10, "Failed_Checks": 0, "Skipped_Checks": 4, } ... } ``` #### [optional] `?test_session_token=` #### [optional] `X-Datadog-Test-Session-Token` #### [optional] `?return_all=` ``` curl -X GET 'http://0.0.0.0:8126/test/trace_check/summary' ``` ### /test/session/integrations (PUT) Update information about the current tested integration. #### [optional] `?test_session_token=` #### [optional] `X-Datadog-Test-Session-Token` ``` curl -X PUT 'http://0.0.0.0:8126/test/session/integrations' -d '{"integration_name": [INTEGRATION_NAME], "integration_version": [INTEGRATION_VERSION], "dependency_name": [DEPENDENCY_NAME], "tracer_language": [TRACER_LANGUAGE], "tracer_version": [TRACER_VERSION]}' ``` ### /test/integrations/tested_versions (GET) Return a csv list of all tested integrations received by the agent. The format of returned data will be: `tracer_language,tracer_version,integration_name,integration_version,dependency_name`. #### [optional] `?test_session_token=` #### [optional] `X-Datadog-Test-Session-Token` ``` curl -X GET 'http://0.0.0.0:8126/test/integrations/tested_versions' ``` ### /v0.1/pipeline_stats Mimics the pipeline_stats endpoint of the agent, but always returns OK, and logs a line everytime it's called. ### /v1/logs (HTTP) Accepts OpenTelemetry Protocol (OTLP) v1.7.0 logs in protobuf format via HTTP. This endpoint validates and decodes OTLP logs payloads for testing OpenTelemetry logs exporters and libraries. The HTTP endpoint accepts `POST` requests with `Content-Type: application/x-protobuf` and `Content-Type: application/json` and stores the decoded logs for retrieval via the `/test/session/logs` endpoint. ### /v1/metrics (HTTP) Accepts OpenTelemetry Protocol (OTLP) v1.7.0 metrics in protobuf format via HTTP. This endpoint validates and decodes OTLP metrics payloads for testing OpenTelemetry metrics exporters and libraries. The HTTP endpoint accepts `POST` requests with `Content-Type: application/x-protobuf` and `Content-Type: application/json` and stores the decoded metrics for retrieval via the `/test/session/metrics` endpoint. ### OTLP Logs and Metrics via GRPC OTLP logs and metrics can also be sent via GRPC using the OpenTelemetry `LogsService.Export` and `MetricsService.Export` methods respectively. The GRPC server implements the standard OTLP service interfaces and forwards all requests to the HTTP server, ensuring consistent processing and session management. **Note:** OTLP endpoints are served on separate ports from the main APM endpoints (default: 8126): - **HTTP**: Port 4318 (default) - Use `--otlp-http-port` to configure - **GRPC**: Port 4317 (default) - Use `--otlp-grpc-port` to configure Both protocols store decoded data for retrieval via the `/test/session/logs` and `/test/session/metrics` HTTP endpoints respectively. GRPC Client → GRPC Server → HTTP POST → HTTP Server → Agent Storage ↓ ↓ (forwards protobuf) (session management) ↓ ↓ HTTP Retrievable via Response /test/session/{logs,metrics} ### /tracer_flare/v1 Mimics the tracer_flare endpoint of the agent. Returns OK if the flare contains the required form fields, otherwise `400`. Logs a line everytime it's called and stores the tracer flare details in the request under `"_tracer_flare"`. ### /test/session/tracerflares Return all tracer-flares that have been received by the agent for the given session token. #### [optional] `?test_session_token=` #### [optional] `X-Datadog-Test-Session-Token` Returns the tracer-flares in the following json format: ```json [ { "source": "...", "case_id": "...", "email": "...", "hostname": "...", "flare_file": "...", } ] ``` `flare_file` is the base64 encoded content of the tracer-flare payload. If there was an error parsing the tracer-flare form, that will be recorded under `error`. ### /test/settings (POST) Allows to change some settings on the fly. This endpoint takes a POST request with a json content listing the keys and values to apply. ```js { 'key': value } ``` Supported keys: - `trace_request_delay`: sets a delay to apply to trace and telemetry requests ``` curl -X POST 'http://0.0.0.0:8126/test/settings' -d '{ "trace_request_delay": 5 }' ``` ## Development ### Prerequisites A Python version of 3.8 or above and [`riot`](https://github.com/Datadog/riot) are required. It is recommended to create and work out of a virtualenv: python3.12 -m venv .venv source .venv/bin/activate pip install -e '.[testing]' ### Running the tests To run the tests (in Python 3.12): riot run -p3.12 test Note: if snapshots need to be (re)generated in the tests set the environment variable `GENERATE_SNAPSHOTS=1`. GENERATE_SNAPSHOTS=1 riot run --pass-env -p3.12 test -k test_trace_missing_received ### Linting and formatting To lint, format and type-check the code: riot run -s flake8 riot run -s fmt riot run -s mypy ### Docker To build (and tag) the dockerfile: ```bash docker build --tag testagent . ``` Run the tagged image: ```bash docker run --rm -v ${PWD}/snaps:/snapshots --publish 8126:8126 agent ``` ### Release notes This project follows [`semver`](https://semver.org/) and so bug fixes, breaking changes, new features, etc must be accompanied by a release note. To generate a release note: ```bash riot run reno new <short-description-of-change> ``` document the changes in the generated file, remove the irrelevant sections and commit the release note with the change. ### Releasing 1. Checkout the `master` branch and make sure it's up to date. ```bash git checkout master && git pull ``` 2. Generate the release notes and use [`pandoc`](https://pandoc.org/) to format them for Github: ```bash riot run -s reno report --no-show-source | pandoc -f rst -t gfm --wrap=none ``` Copy the output into a new release: https://github.com/DataDog/dd-apm-test-agent/releases/new. 2. Enter a tag for the release (following [`semver`](https://semver.org)) (eg. `v1.1.3`, `v1.0.3`, `v1.2.0`). 3. Use the tag without the `v` as the title. 4. Save the release as a draft and pass the link to someone else to give a quick review. 5. If all looks good hit publish
text/markdown
Kyle Verhoog
kyle@verhoog.ca
null
null
BSD 3
null
[ "Programming Language :: Python", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13" ]
[]
https://github.com/Datadog/dd-apm-test-agent
null
>=3.8
[]
[]
[]
[ "aiohttp", "ddsketch[serialization]", "msgpack", "requests", "typing_extensions", "yarl", "requests-aws4auth", "jinja2>=3.0.0", "pyyaml", "opentelemetry-proto<1.37.0,>1.33.0", "protobuf>=3.19.0", "grpcio<2.0,>=1.66.2", "pywin32; sys_platform == \"win32\"", "ddtrace==3.15.0; extra == \"testing\"", "pytest; extra == \"testing\"", "riot==0.20.1; extra == \"testing\"", "PyYAML==6.0.3; extra == \"testing\"" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.12.9
2026-02-21T00:29:56.737399
ddapm_test_agent-1.42.0.tar.gz
1,228,659
99/6f/4878efa9d596b5caebd7008073e1a77a714683a3ba5743e841d200106816/ddapm_test_agent-1.42.0.tar.gz
source
sdist
null
false
59b62c9be690c038c6b8ffdc8db65110
ea0819655799c5f9220a6526df2c6937a50f491f5022590d3ab40f5cfcef5a5a
996f4878efa9d596b5caebd7008073e1a77a714683a3ba5743e841d200106816
null
[ "LICENSE.BSD3", "LICENSE.apache2" ]
257
2.4
uipath-langchain
0.5.78
Python SDK that enables developers to build and deploy LangGraph agents to the UiPath Cloud Platform
# UiPath LangChain Python SDK [![PyPI - Version](https://img.shields.io/pypi/v/uipath-langchain)](https://pypi.org/project/uipath-langchain/) [![PyPI downloads](https://img.shields.io/pypi/dm/uipath-langchain.svg)](https://pypi.org/project/uipath-langchain/) [![Python versions](https://img.shields.io/pypi/pyversions/uipath-langchain.svg)](https://pypi.org/project/uipath-langchain/) A Python SDK that enables developers to build and deploy LangGraph agents to the UiPath Cloud Platform. It provides programmatic interaction with UiPath Cloud Platform services and human-in-the-loop (HITL) semantics through Action Center integration. This package is an extension to the [UiPath Python SDK](https://github.com/UiPath/uipath-python) and implements the [UiPath Runtime Protocol](https://github.com/UiPath/uipath-runtime-python). This [quickstart guide](https://uipath.github.io/uipath-python/) walks you through deploying your first agent to UiPath Cloud Platform. Check out these [sample projects](https://github.com/UiPath/uipath-langchain-python/tree/main/samples) to see the SDK in action. ## Requirements - Python 3.11 or higher - UiPath Automation Cloud account ## Installation ```bash pip install uipath-langchain ``` using `uv`: ```bash uv add uipath-langchain ``` ## Configuration ### Environment Variables Create a `.env` file in your project root with the following variables: ``` UIPATH_URL=https://cloud.uipath.com/ACCOUNT_NAME/TENANT_NAME UIPATH_ACCESS_TOKEN=YOUR_TOKEN_HERE ``` ## Command Line Interface (CLI) The SDK provides a command-line interface for creating, packaging, and deploying LangGraph Agents: ### Initialize a Project ```bash uipath init ``` Running `uipath init` will process the graph definitions in the `langgraph.json` file and create the corresponding `entry-points.json` file needed for deployment. For more details on the configuration format, see the [UiPath configuration specifications](https://github.com/UiPath/uipath-python/blob/main/specs/README.md). ### Authentication ```bash uipath auth ``` This command opens a browser for authentication and creates/updates your `.env` file with the proper credentials. ### Debug a Project ```bash uipath run GRAPH [INPUT] ``` Executes the agent with the provided JSON input arguments. ### Package a Project ```bash uipath pack ``` Packages your project into a `.nupkg` file that can be deployed to UiPath. **Note:** Your `pyproject.toml` must include: - A description field (avoid characters: &, <, >, ", ', ;) - Author information Example: ```toml description = "Your package description" authors = [{name = "Your Name", email = "your.email@example.com"}] ``` ### Publish a Package ```bash uipath publish ``` Publishes the most recently created package to your UiPath Orchestrator. ## Project Structure To properly use the CLI for packaging and publishing, your project should include: - A `pyproject.toml` file with project metadata - A `langgraph.json` file with your graph definitions (e.g., `"graphs": {"agent": "graph.py:graph"}`) - A `entry-points.json` file (generated by `uipath init`) - A `bindings.json` file (generated by `uipath init`) to configure resource overrides - Any Python files needed for your automation ## Development ### Developer Tools Check out [uipath-dev](https://github.com/uipath/uipath-dev-python) - an interactive terminal application for building, testing, and debugging UiPath Python runtimes, agents, and automation scripts. ### Setting Up a Development Environment Please read our [contribution guidelines](https://github.com/UiPath/uipath-langchain-python/blob/main/CONTRIBUTING.md) before submitting a pull request. ### Special Thanks A huge thank-you to the open-source community and the maintainers of the libraries that make this project possible: - [LangChain](https://github.com/langchain-ai/langchain) for providing a powerful framework for building stateful LLM applications. - [Pydantic](https://github.com/pydantic/pydantic) for reliable, typed configuration and validation. - [OpenInference](https://github.com/Arize-ai/openinference) for observability and instrumentation support.
text/markdown
null
null
null
Marius Cosareanu <marius.cosareanu@uipath.com>, Cristian Pufu <cristian.pufu@uipath.com>
null
null
[ "Intended Audience :: Developers", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Software Development :: Build Tools" ]
[]
null
null
>=3.11
[]
[]
[]
[ "httpx>=0.27.0", "jsonpath-ng>=1.7.0", "jsonschema-pydantic-converter>=0.1.9", "langchain-core<2.0.0,>=1.2.11", "langchain-mcp-adapters==0.2.1", "langchain-openai<2.0.0,>=1.0.0", "langchain<2.0.0,>=1.0.0", "langgraph-checkpoint-sqlite<4.0.0,>=3.0.3", "langgraph<2.0.0,>=1.0.0", "mcp==1.26.0", "openinference-instrumentation-langchain>=0.1.56", "pydantic-settings>=2.6.0", "python-dotenv>=1.0.1", "uipath-runtime<0.10.0,>=0.9.0", "uipath<2.9.0,>=2.8.46", "boto3-stubs>=1.41.4; extra == \"bedrock\"", "langchain-aws>=0.2.35; extra == \"bedrock\"", "google-generativeai>=0.8.0; extra == \"vertex\"", "langchain-google-genai>=2.0.0; extra == \"vertex\"" ]
[]
[]
[]
[ "Homepage, https://uipath.com", "Repository, https://github.com/UiPath/uipath-langchain-python", "Documentation, https://uipath.github.io/uipath-python/" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:29:30.706657
uipath_langchain-0.5.78.tar.gz
9,029,102
3a/02/9d5e0470fce75b3eb4a28f29cb07be1f057cf99e1c0844998eeb69539801/uipath_langchain-0.5.78.tar.gz
source
sdist
null
false
d51e0198e457e6f85eb30245ef8f8b94
5a409066831131f572462827b63c60f16ac523b4d6b3b8add2bad43702f255d8
3a029d5e0470fce75b3eb4a28f29cb07be1f057cf99e1c0844998eeb69539801
null
[ "LICENSE" ]
3,340
2.4
bioblend
1.8.0
Library for interacting with the Galaxy API
.. image:: https://img.shields.io/pypi/v/bioblend.svg :target: https://pypi.org/project/bioblend/ :alt: latest version available on PyPI .. image:: https://readthedocs.org/projects/bioblend/badge/ :alt: Documentation Status :target: https://bioblend.readthedocs.io/ .. image:: https://badges.gitter.im/galaxyproject/bioblend.svg :alt: Join the chat at https://gitter.im/galaxyproject/bioblend :target: https://gitter.im/galaxyproject/bioblend?utm_source=badge&utm_medium=badge&utm_campaign=pr-badge&utm_content=badge BioBlend is a Python library for interacting with the `Galaxy`_ API. BioBlend is supported and tested on: - Python 3.10 - 3.14 - Galaxy release 19.05 and later. Full docs are available at https://bioblend.readthedocs.io/ with a quick library overview also available in `ABOUT.rst <./ABOUT.rst>`_. .. References/hyperlinks used above .. _Galaxy: https://galaxyproject.org/
text/x-rst
null
Enis Afgan <afgane@gmail.com>
null
Nicola Soranzo <nicola.soranzo@earlham.ac.uk>
null
null
[ "Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Topic :: Scientific/Engineering :: Bio-Informatics", "Typing :: Typed" ]
[]
null
null
>=3.10
[]
[]
[]
[ "PyYAML", "requests>=2.20.0", "requests-toolbelt!=0.9.0,>=0.5.1", "tuspy", "pytest; extra == \"testing\"" ]
[]
[]
[]
[ "Homepage, https://bioblend.readthedocs.io/", "Documentation, https://bioblend.readthedocs.io/", "Bug Tracker, https://github.com/galaxyproject/bioblend/issues", "Source Code, https://github.com/galaxyproject/bioblend" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:29:07.616803
bioblend-1.8.0.tar.gz
159,367
44/e1/5d7c8664538efc02324de2e98a93c04014f7f9839ffa295aa4b7f98a172c/bioblend-1.8.0.tar.gz
source
sdist
null
false
a010d032e86f75af1cffb76650bb472c
6dbb2cbfa243f52d3bbb7107425af535f2ee9fbb7a50bfdc496c068a4a474fbd
44e15d7c8664538efc02324de2e98a93c04014f7f9839ffa295aa4b7f98a172c
MIT
[ "LICENSE" ]
1,533
2.4
TatSu
5.17.1
TatSu takes a grammar in a variation of EBNF as input, and outputs a memoizing PEG/Packrat parser in Python.
.. Copyright (c) 2017-2026 Juancarlo Añez (apalala@gmail.com) .. SPDX-License-Identifier: BSD-4-Clause .. |dragon| unicode:: 0x7ADC .. unicode dragon .. |nbsp| unicode:: 0xA0 .. non breakable space .. |TatSu| replace:: |dragon|\ |nbsp|\ **TatSu** .. |TatSu-LTS| replace:: |dragon|\ |nbsp|\ **TatSu-LTS** .. _RELEASES: https://github.com/neogeny/TatSu/releases | |license| | |pyversions| | |fury| | |actions| | |docs| | |installs| | |sponsor| | *At least for the people who send me mail about a new language that they're designing, the general advice is: do it to learn about how to write a compiler. Don't have any expectations that anyone will use it, unless you hook up with some sort of organization in a position to push it hard. It's a lottery, and some can buy a lot of the tickets. There are plenty of beautiful languages (more beautiful than C) that didn't catch on. But someone does win the lottery, and doing a language at least teaches you something.* `Dennis Ritchie`_ (1941-2011) Creator of the C_ programming language and of Unix_ |TatSu| ======= |TatSu| is a tool that takes grammars in extended `EBNF`_ as input, and outputs `memoizing`_ (`Packrat`_) `PEG`_ parsers in `Python`_. The classic variations of EBNF_ (Tomassetti, EasyExtend, Wirth) and `ISO EBNF`_ are also supported as input grammar format. Why use a PEG_ parser? Because `regular languages`_ (those parseable with Python's ``re`` package) *"cannot count"*. Any language with nested structures or with balancing of demarcations requires more than regular expressions to be parsed. |TatSu| can compile a grammar stored in a string into a ``tatsu.grammars.Grammar`` object that can be used to parse any given input, much like the `re`_ module does with regular expressions, or it can generate a Python_ module that implements the parser. |TatSu| supports `left-recursive`_ rules in PEG_ grammars using the algorithm_ by *Laurent* and *Mens*. The generated AST_ has the expected left associativity. |TatSu| expects a maintained version of Python (>=3.14 at the moment). While no code in |TatSu| yet depends on new language or standard library features, the authors don't want to be constrained by Python version compatibility considerations when developing future releases. That said, currently all tests run in versions down to Python 3.12. *If you need support for previous versions of Python, please consider* `TatSu-LTS`_, *a friendly fork of* |TatSu| *aimed at compatibility with other versions of Python still used by many projects. The developers of both projects work together to promote compatibility with most versions of Python.* .. _algorithm: http://norswap.com/pubs/sle2016.pdf .. _TatSu-LTS: https://pypi.org/project/TatSu-LTS/ Installation ------------ .. code-block:: bash $ pip install TatSu Using the Tool -------------- |TatSu| can be used as a library, much like `Python`_'s ``re``, by embedding grammars as strings and generating grammar models instead of generating Python_ code. This compiles the grammar and generates an in-memory *parser* that can subsequently be used for parsing input with: .. code-block:: python parser = tatsu.compile(grammar) Compiles the grammar and parses the given input producing an AST_ as result: .. code-block:: python ast = tatsu.parse(grammar, input) The result is equivalent to calling: .. code-block:: python parser = compile(grammar) ast = parser.parse(input) Compiled grammars are cached for efficiency. This compiles the grammar to the `Python`_ source code that implements the parser: .. code-block:: python parser_source = tatsu.to_python_sourcecode(grammar) This is an example of how to use |TatSu| as a library: .. code-block:: python GRAMMAR = ''' @@grammar::CALC start = expression $ ; expression = | expression '+' term | expression '-' term | term ; term = | term '*' factor | term '/' factor | factor ; factor = | '(' expression ')' | number ; number = /\d+/ ; ''' if __name__ == '__main__': import json from tatsu import parse from tatsu.util import asjson ast = parse(GRAMMAR, '3 + 5 * ( 10 - 20 )') print(json.dumps(asjson(ast), indent=2)) .. |TatSu| will use the first rule defined in the grammar as the *start* rule. This is the output: .. code-block:: console [ "3", "+", [ "5", "*", [ "10", "-", "20" ] ] ] Documentation ------------- For a detailed explanation of what |TatSu| is capable of, please see the documentation_. .. _documentation: http://tatsu.readthedocs.io/ Questions? ---------- Please use the `[tatsu]`_ tag on `StackOverflow`_ for general Q&A, and limit GitHub issues to bugs, enhancement proposals, and feature requests. .. _[tatsu]: https://stackoverflow.com/tags/tatsu/info Changes ------- See the `RELEASES`_ for details. License ------- You may use |TatSu| under the terms of the `BSD`_-style license described in the enclosed `LICENSE`_ file. *If your project requires different licensing* please `email`_. For Fun ------- This is a diagram of the grammar for |TatSu|'s own grammar language: .. code:: console start ●─grammar─■ grammar∷Grammar ●─ [title](`TATSU`)──┬→───────────────────────────────────┬── [`rules`]+(rule)──┬→───────────────────────────────┬──⇥ ␃ ─■ ├→──┬─ [`directives`]+(directive)─┬──┤ ├→──┬─ [`rules`]+(rule)───────┬──┤ │ └─ [`keywords`]+(keyword)─────┘ │ │ └─ [`keywords`]+(keyword)─┘ │ └───────────────────────────────────<┘ └───────────────────────────────<┘ directive ●─'@@'─ !['keyword'] ✂ ───┬─ [name](──┬─'comments'─────┬─) ✂ ─ ✂ ─'::' ✂ ─ [value](regex)────────┬─ ✂ ──■ │ └─'eol_comments'─┘ │ ├─ [name]('whitespace') ✂ ─'::' ✂ ─ [value](──┬─regex───┬─)────────────┤ │ ├─string──┤ │ │ ├─'None'──┤ │ │ ├─'False'─┤ │ │ └─`None`──┘ │ ├─ [name](──┬─'nameguard'──────┬─) ✂ ───┬─'::' ✂ ─ [value](boolean)─┬──┤ │ ├─'ignorecase'─────┤ └─ [value](`True`)──────────┘ │ │ ├─'left_recursion'─┤ │ │ ├─'parseinfo'──────┤ │ │ └─'memoization'────┘ │ ├─ [name]('grammar') ✂ ─'::' ✂ ─ [value](word)─────────────────────────┤ └─ [name]('namechars') ✂ ─'::' ✂ ─ [value](string)─────────────────────┘ keywords ●───┬─keywords─┬───■ └─────────<┘ keyword ●─'@@keyword' ✂ ─'::' ✂ ───┬→──────────────────────────────────┬───■ ├→ @+(──┬─word───┬─)─ ![──┬─':'─┬─]─┤ │ └─string─┘ └─'='─┘ │ └──────────────────────────────────<┘ paramdef ●───┬─'::' ✂ ─ [params](params)──────────────────────────────────────┬──■ └─'(' ✂ ───┬─ [kwparams](kwparams)─────────────────────────┬─')'─┘ ├─ [params](params)',' ✂ ─ [kwparams](kwparams)─┤ └─ [params](params)─────────────────────────────┘ rule∷Rule ●─ [decorators](──┬→──────────┬──) [name](name) ✂ ───┬─→ >(paramdef) ─┬───┬─→'<' ✂ ─ [base](known_name)─┬───┬─'='──┬─ ✂ ─ [exp](expre)RULE_END ✂ ──■ ├→decorator─┤ └─→──────────────┘ └─→───────────────────────────┘ ├─':='─┤ └──────────<┘ └─':'──┘ RULE_END ●───┬─EMPTYLINE──┬─→';'─┬──┬──■ │ └─→────┘ │ ├─⇥ ␃ │ └─';'──────────────────┘ EMPTYLINE ●─/(?:\s*(?:\r?\n|\r)){2,}/──■ decorator ●─'@'─ !['@'] ✂ ─ @(──┬─'override'─┬─)─■ ├─'name'─────┤ └─'nomemo'───┘ params ●─ @+(first_param)──┬→────────────────────────────┬───■ ├→',' @+(literal)─ !['='] ✂ ──┤ └────────────────────────────<┘ first_param ●───┬─path────┬──■ └─literal─┘ kwparams ●───┬→────────────┬───■ ├→',' ✂ ─pair─┤ └────────────<┘ pair ●─ @+(word)'=' ✂ ─ @+(literal)─■ expre ●───┬─choice───┬──■ └─sequence─┘ choice∷Choice ●───┬─→'|' ✂ ──┬─ @+(option)──┬─'|' ✂ ─ @+(option)─┬───■ └─→────────┘ └───────────────────<┘ option∷Option ●─ @(sequence)─■ sequence∷Sequence ●───┬── &[element',']──┬→───────────────┬───┬──■ │ ├→',' ✂ ─element─┤ │ │ └───────────────<┘ │ └───┬── ![EMPTYLINE]element─┬───────────┘ └──────────────────────<┘ element ●───┬─rule_include─┬──■ ├─named────────┤ ├─override─────┤ └─term─────────┘ rule_include∷RuleInclude ●─'>' ✂ ─ @(known_name)─■ named ●───┬─named_list───┬──■ └─named_single─┘ named_list∷NamedList ●─ [name](name)'+:' ✂ ─ [exp](term)─■ named_single∷Named ●─ [name](name)':' ✂ ─ [exp](term)─■ override ●───┬─override_list──────────────┬──■ ├─override_single────────────┤ └─override_single_deprecated─┘ override_list∷OverrideList ●─'@+:' ✂ ─ @(term)─■ override_single∷Override ●─'@:' ✂ ─ @(term)─■ override_single_deprecated∷Override ●─'@' ✂ ─ @(term)─■ term ●───┬─void───────────────┬──■ ├─gather─────────────┤ ├─join───────────────┤ ├─left_join──────────┤ ├─right_join─────────┤ ├─empty_closure──────┤ ├─positive_closure───┤ ├─closure────────────┤ ├─optional───────────┤ ├─skip_to────────────┤ ├─lookahead──────────┤ ├─negative_lookahead─┤ ├─cut────────────────┤ ├─cut_deprecated─────┤ └─atom───────────────┘ group∷Group ●─'(' ✂ ─ @(expre)')' ✂ ──■ gather ●── &[atom'.{'] ✂ ───┬─positive_gather─┬──■ └─normal_gather───┘ positive_gather∷PositiveGather ●─ [sep](atom)'.{' [exp](expre)'}'──┬─'+'─┬─ ✂ ──■ └─'-'─┘ normal_gather∷Gather ●─ [sep](atom)'.{' ✂ ─ [exp](expre)'}'──┬─→'*' ✂ ──┬─ ✂ ──■ └─→────────┘ join ●── &[atom'%{'] ✂ ───┬─positive_join─┬──■ └─normal_join───┘ positive_join∷PositiveJoin ●─ [sep](atom)'%{' [exp](expre)'}'──┬─'+'─┬─ ✂ ──■ └─'-'─┘ normal_join∷Join ●─ [sep](atom)'%{' ✂ ─ [exp](expre)'}'──┬─→'*' ✂ ──┬─ ✂ ──■ └─→────────┘ left_join∷LeftJoin ●─ [sep](atom)'<{' ✂ ─ [exp](expre)'}'──┬─'+'─┬─ ✂ ──■ └─'-'─┘ right_join∷RightJoin ●─ [sep](atom)'>{' ✂ ─ [exp](expre)'}'──┬─'+'─┬─ ✂ ──■ └─'-'─┘ positive_closure∷PositiveClosure ●───┬─'{' @(expre)'}'──┬─'-'─┬─ ✂ ──┬──■ │ └─'+'─┘ │ └─ @(atom)'+' ✂ ────────────────┘ closure∷Closure ●───┬─'{' @(expre)'}'──┬─→'*'─┬─ ✂ ──┬──■ │ └─→────┘ │ └─ @(atom)'*' ✂ ─────────────────┘ empty_closure∷EmptyClosure ●─'{}' ✂ ─ @( ∅ )─■ optional∷Optional ●───┬─'[' ✂ ─ @(expre)']' ✂ ──────────┬──■ └─ @(atom)─ ![──┬─'?"'─┬─]'?' ✂ ──┘ ├─"?'"─┤ └─'?/'─┘ lookahead∷Lookahead ●─'&' ✂ ─ @(term)─■ negative_lookahead∷NegativeLookahead ●─'!' ✂ ─ @(term)─■ skip_to∷SkipTo ●─'->' ✂ ─ @(term)─■ atom ●───┬─group────┬──■ ├─token────┤ ├─alert────┤ ├─constant─┤ ├─call─────┤ ├─pattern──┤ ├─dot──────┤ └─eof──────┘ call∷Call ●─word─■ void∷Void ●─'()' ✂ ──■ fail∷Fail ●─'!()' ✂ ──■ cut∷Cut ●─'~' ✂ ──■ cut_deprecated∷Cut ●─'>>' ✂ ──■ known_name ●─name ✂ ──■ name ●─word─■ constant∷Constant ●── &['`']──┬─/(?ms)```((?:.|\n)*?)```/──┬──■ ├─'`' @(literal)'`'──────────┤ └─/`(.*?)`/──────────────────┘ alert∷Alert ●─ [level](/\^+/─) [message](constant)─■ token∷Token ●───┬─string─────┬──■ └─raw_string─┘ literal ●───┬─string─────┬──■ ├─raw_string─┤ ├─boolean────┤ ├─word───────┤ ├─hex────────┤ ├─float──────┤ ├─int────────┤ └─null───────┘ string ●─STRING─■ raw_string ●─//─ @(STRING)─■ STRING ●───┬─ @(/"((?:[^"\n]|\\"|\\\\)*?)"/─) ✂ ─────┬──■ └─ @(/r"'((?:[^'\n]|\\'|\\\\)*?)'"/─) ✂ ──┘ hex ●─/0[xX](?:\d|[a-fA-F])+/──■ float ●─/[-+]?(?:\d+\.\d*|\d*\.\d+)(?:[Ee][-+]?\d+)?/──■ int ●─/[-+]?\d+/──■ path ●─/(?!\d)\w+(?:::(?!\d)\w+)+/──■ word ●─/(?!\d)\w+/──■ dot∷Dot ●─'/./'─■ pattern∷Pattern ●─regexes─■ regexes ●───┬→─────────────┬───■ ├→'+' ✂ ─regex─┤ └─────────────<┘ regex ●───┬─'/' ✂ ─ @(/(?:[^/\\]|\\/|\\.)*/─)'/' ✂ ──┬──■ ├─'?' @(STRING)────────────────────────────┤ └─deprecated_regex─────────────────────────┘ deprecated_regex ●─'?/' ✂ ─ @(/(?:.|\n)*?(?=/\?)/─)//\?+/─ ✂ ──■ boolean ●───┬─'True'──┬──■ └─'False'─┘ null ●─'None'─■ eof∷EOF ●─'$' ✂ ──■ .. _ANTLR: http://www.antlr.org/ .. _AST: http://en.wikipedia.org/wiki/Abstract_syntax_tree .. _Abstract Syntax Tree: http://en.wikipedia.org/wiki/Abstract_syntax_tree .. _Algol W: http://en.wikipedia.org/wiki/Algol_W .. _Algorithms + Data Structures = Programs: http://www.amazon.com/Algorithms-Structures-Prentice-Hall-Automatic-Computation/dp/0130224189/ .. _BSD: http://en.wikipedia.org/wiki/BSD_licenses#2-clause_license_.28.22Simplified_BSD_License.22_or_.22FreeBSD_License.22.29 .. _Basel Shishani: https://bitbucket.org/basel-shishani .. _C: http://en.wikipedia.org/wiki/C_language .. _CHANGELOG: https://github.com/neogeny/TatSu/releases .. _CSAIL at MIT: http://www.csail.mit.edu/ .. _Cyclomatic complexity: http://en.wikipedia.org/wiki/Cyclomatic_complexity .. _David Röthlisberger: https://bitbucket.org/drothlis/ .. _Dennis Ritchie: http://en.wikipedia.org/wiki/Dennis_Ritchie .. _EBNF: http://en.wikipedia.org/wiki/Ebnf .. _ISO EBNF: http://en.wikipedia.org/wiki/Ebnf .. _English: http://en.wikipedia.org/wiki/English_grammar .. _Euler: http://en.wikipedia.org/wiki/Euler_programming_language .. _Grako: https://bitbucket.org/neogeny/grako/ .. _Jack: http://en.wikipedia.org/wiki/Javacc .. _Japanese: http://en.wikipedia.org/wiki/Japanese_grammar .. _KLOC: http://en.wikipedia.org/wiki/KLOC .. _Kathryn Long: https://bitbucket.org/starkat .. _Keywords: https://en.wikipedia.org/wiki/Reserved_word .. _`left-recursive`: https://en.wikipedia.org/wiki/Left_recursion .. _LL(1): http://en.wikipedia.org/wiki/LL(1) .. _Marcus Brinkmann: http://blog.marcus-brinkmann.de/ .. _MediaWiki: http://www.mediawiki.org/wiki/MediaWiki .. _Modula-2: http://en.wikipedia.org/wiki/Modula-2 .. _Modula: http://en.wikipedia.org/wiki/Modula .. _Oberon-2: http://en.wikipedia.org/wiki/Oberon-2 .. _Oberon: http://en.wikipedia.org/wiki/Oberon_(programming_language) .. _PEG and Packrat parsing mailing list: https://lists.csail.mit.edu/mailman/listinfo/peg .. _PEG.js: http://pegjs.majda.cz/ .. _PEG: http://en.wikipedia.org/wiki/Parsing_expression_grammar .. _PL/0: http://en.wikipedia.org/wiki/PL/0 .. _Packrat: http://bford.info/packrat/ .. _Pascal: http://en.wikipedia.org/wiki/Pascal_programming_language .. _Paul Sargent: https://bitbucket.org/PaulS/ .. _Perl: http://www.perl.org/ .. _PyPy team: http://pypy.org/people.html .. _PyPy: http://pypy.org/ .. _Python Weekly: http://www.pythonweekly.com/ .. _Python: http://python.org .. _Reserved Words: https://en.wikipedia.org/wiki/Reserved_word .. _Robert Speer: https://bitbucket.org/r_speer .. _Ruby: http://www.ruby-lang.org/ .. _Semantic Graph: http://en.wikipedia.org/wiki/Abstract_semantic_graph .. _StackOverflow: http://stackoverflow.com/tags/tatsu/info .. _Sublime Text: https://www.sublimetext.com .. _TatSu Forum: https://groups.google.com/forum/?fromgroups#!forum/tatsu .. _UCAB: http://www.ucab.edu.ve/ .. _USB: http://www.usb.ve/ .. _Unix: http://en.wikipedia.org/wiki/Unix .. _VIM: http://www.vim.org/ .. _WTK: http://en.wikipedia.org/wiki/Well-known_text .. _Warth et al: http://www.vpri.org/pdf/tr2007002_packrat.pdf .. _Well-known text: http://en.wikipedia.org/wiki/Well-known_text .. _Wirth: http://en.wikipedia.org/wiki/Niklaus_Wirth .. _`LICENSE`: LICENSE .. _basel-shishani: https://bitbucket.org/basel-shishani .. _blog post: http://dietbuddha.blogspot.com/2012/12/52python-encapsulating-exceptions-with.html .. _colorama: https://pypi.python.org/pypi/colorama/ .. _context managers: http://docs.python.org/2/library/contextlib.html .. _declensions: http://en.wikipedia.org/wiki/Declension .. _drothlis: https://bitbucket.org/drothlis .. _email: mailto:apalala@gmail.com .. _exceptions: http://www.jeffknupp.com/blog/2013/02/06/write-cleaner-python-use-exceptions/ .. _franz\_g: https://bitbucket.org/franz_g .. _gapag: https://bitbucket.org/gapag .. _gegenschall: https://bitbucket.org/gegenschall .. _gkimbar: https://bitbucket.org/gkimbar .. _introduced: http://dl.acm.org/citation.cfm?id=964001.964011 .. _jimon: https://bitbucket.org/jimon .. _keyword: https://en.wikipedia.org/wiki/Reserved_word .. _keywords: https://en.wikipedia.org/wiki/Reserved_word .. _lambdafu: http://blog.marcus-brinkmann.de/ .. _leewz: https://bitbucket.org/leewz .. _linkdd: https://bitbucket.org/linkdd .. _make a donation: https://www.paypal.com/cgi-bin/webscr?cmd=_s-xclick&hosted_button_id=P9PV7ZACB669J .. _memoizing: http://en.wikipedia.org/wiki/Memoization .. _nehz: https://bitbucket.org/nehz .. _neumond: https://bitbucket.org/neumond .. _parsewkt: https://github.com/cleder/parsewkt .. _pauls: https://bitbucket.org/pauls .. _pgebhard: https://bitbucket.org/pgebhard .. _r\_speer: https://bitbucket.org/r_speer .. _raw string literal: https://docs.python.org/3/reference/lexical_analysis.html#string-and-bytes-literals .. _re: https://docs.python.org/3.7/library/re.html .. _regular languages: https://en.wikipedia.org/wiki/Regular_language .. _regex: https://pypi.python.org/pypi/regex .. _siemer: https://bitbucket.org/siemer .. _sjbrownBitbucket: https://bitbucket.org/sjbrownBitbucket .. _smc.mw: https://github.com/lambdafu/smc.mw .. _starkat: https://bitbucket.org/starkat .. _tonico\_strasser: https://bitbucket.org/tonico_strasser .. _vinay.sajip: https://bitbucket.org/vinay.sajip .. _vmuriart: https://bitbucket.org/vmuriart .. |fury| image:: https://badge.fury.io/py/TatSu.svg :target: https://badge.fury.io/py/TatSu .. |license| image:: https://img.shields.io/badge/license-BSD-blue.svg :target: https://raw.githubusercontent.com/neogeny/tatsu/master/LICENSE .. |pyversions| image:: https://img.shields.io/pypi/pyversions/tatsu.svg :target: https://pypi.python.org/pypi/tatsu .. |actions| image:: https://github.com/neogeny/TatSu/actions/workflows/default.yml/badge.svg :target: https://github.com/neogeny/TatSu/actions/workflows/default.yml .. |docs| image:: https://readthedocs.org/projects/tatsu/badge/?version=stable&logo=readthedocs :target: http://tatsu.readthedocs.io/en/stable/ .. |installs| image:: https://img.shields.io/pypi/dm/tatsu.svg?label=installs&logo=pypi :target: https://pypistats.org/packages/tatsu .. |downloads| image:: https://img.shields.io/github/downloads/neogeny/tatsu/total?label=downloads :target: https://pypistats.org/packages/tatsu .. |sponsor| image:: https://img.shields.io/badge/Sponsor-EA4AAA?label=TatSu :target: https://github.com/sponsors/neogeny
text/x-rst
null
Juancarlo Añez <apalala@gmail.com>
null
null
SPDX-License-Identifier: BSD-4-Clause TATSU - A PEG/Packrat parser generator for Python Copyright (c) 2017-2026 Juancarlo Añez (apalala@gmail.com) All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: 1. Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. 2. Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. 3. All advertising materials mentioning features or use of this software must display the following acknowledgement: This product includes software developed by Juancarlo Añez (https://github.com/apalala). 4. Neither the name of the copyright holder nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDER AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
EBNF, PEG, grammar, packrat, parser
[ "Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "Natural Language :: English", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Programming Language :: Python :: 3.15", "Topic :: Software Development :: Compilers", "Topic :: Text Processing :: General" ]
[]
null
null
>=3.12
[]
[]
[]
[ "colorama; extra == \"colorization\"", "rich; extra == \"parproc\"" ]
[]
[]
[]
[ "Homepage, https://github.com/neogeny/TatSu", "Repository, https://github.com/neogeny/TatSu", "Documentation, https://tatsu.readthedocs.io/en/stable/" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:29:03.090767
tatsu-5.17.1.tar.gz
313,266
36/78/7a5b45eba65aa101f446caed391d2141d1f0267e501cf1f13cfab3e98780/tatsu-5.17.1.tar.gz
source
sdist
null
false
9cb0e19c6fac006220666cd854fc8bcb
8b11e732206463f31cad57729da6aa60846aa680bdb0c0b64339a8ac7cb6ebb4
36787a5b45eba65aa101f446caed391d2141d1f0267e501cf1f13cfab3e98780
null
[ "LICENSE" ]
0
2.4
clio-kit
2.0.1
CLIO Kit - MCP Servers, Clients, and Tools for AI Agents
# CLIO Kit <!-- mcp-name: io.github.iowarp/adios-mcp --> <!-- mcp-name: io.github.iowarp/arxiv-mcp --> <!-- mcp-name: io.github.iowarp/chronolog-mcp --> <!-- mcp-name: io.github.iowarp/compression-mcp --> <!-- mcp-name: io.github.iowarp/darshan-mcp --> <!-- mcp-name: io.github.iowarp/hdf5-mcp --> <!-- mcp-name: io.github.iowarp/jarvis-mcp --> <!-- mcp-name: io.github.iowarp/lmod-mcp --> <!-- mcp-name: io.github.iowarp/ndp-mcp --> <!-- mcp-name: io.github.iowarp/node-hardware-mcp --> <!-- mcp-name: io.github.iowarp/pandas-mcp --> <!-- mcp-name: io.github.iowarp/parallel-sort-mcp --> <!-- mcp-name: io.github.iowarp/paraview-mcp --> <!-- mcp-name: io.github.iowarp/parquet-mcp --> <!-- mcp-name: io.github.iowarp/plot-mcp --> <!-- mcp-name: io.github.iowarp/slurm-mcp --> [![License: BSD-3-Clause](https://img.shields.io/badge/License-BSD--3--Clause-blue.svg)](https://opensource.org/licenses/BSD-3-Clause) [![PyPI version](https://img.shields.io/pypi/v/clio-kit.svg)](https://pypi.org/project/clio-kit/) [![Python](https://img.shields.io/badge/Python-3.10%2B-blue)](https://www.python.org/) [![FastMCP](https://img.shields.io/badge/FastMCP-2.13%2B-purple)](https://github.com/jlowin/fastmcp) [![CI](https://github.com/iowarp/clio-kit/actions/workflows/quality_control.yml/badge.svg)](https://github.com/iowarp/clio-kit/actions/workflows/quality_control.yml) [![Coverage](https://codecov.io/gh/iowarp/clio-kit/branch/main/graph/badge.svg)](https://codecov.io/gh/iowarp/clio-kit) [![MCP Servers](https://img.shields.io/badge/MCP%20Servers-16-green)](https://github.com/iowarp/clio-kit/tree/main/clio-kit-mcp-servers) [![Ruff](https://img.shields.io/endpoint?url=https://raw.githubusercontent.com/astral-sh/ruff/main/assets/badge/v2.json)](https://github.com/astral-sh/ruff) [![Type Checked](https://img.shields.io/badge/mypy-type%20checked-blue)](http://mypy-lang.org/) [![Package Manager](https://img.shields.io/badge/uv-package%20manager-orange)](https://github.com/astral-sh/uv) [![Security Audit](https://img.shields.io/badge/pip--audit-security%20scanned-green)](https://github.com/pypa/pip-audit) **CLIO Kit** - Part of the IoWarp platform's tooling layer for AI agents. A comprehensive collection of tools, skills, plugins, and extensions. Currently featuring 15+ Model Context Protocol (MCP) servers for scientific computing, with plans to expand to additional agent capabilities. Enables AI agents to interact with HPC resources, scientific data formats, and research datasets. [**Website**](https://toolkit.iowarp.ai/) | [**IOWarp**](https://iowarp.ai) Chat with us on [**Zulip**](https://iowarp.zulipchat.com/#narrow/channel/543872-Agent-Toolkit) or [**join us**](https://iowarp.zulipchat.com/join/e4wh24du356e4y2iw6x6jeay/) Developed by <img src="https://grc.iit.edu/img/logo.png" alt="GRC Logo" width="18" height="18"> [**Gnosis Research Center**](https://grc.iit.edu/) --- ## ❌ Without CLIO Kit Working with scientific data and HPC resources requires manual scripting and tool-specific knowledge: - ❌ Write custom scripts for every HDF5/Parquet file exploration - ❌ Manually craft Slurm job submission scripts - ❌ Switch between multiple tools for data analysis - ❌ No AI assistance for scientific workflows - ❌ Repetitive coding for common research tasks ## ✅ With CLIO Kit AI agents handle scientific computing tasks through natural language: - ✅ **"Analyze the temperature dataset in this HDF5 file"** - HDF5 MCP does it - ✅ **"Submit this simulation to Slurm with 32 cores"** - Slurm MCP handles it - ✅ **"Find papers on neural networks from ArXiv"** - ArXiv MCP searches - ✅ **"Plot the results from this CSV file"** - Plot MCP visualizes - ✅ **"Optimize memory usage for this pandas DataFrame"** - Pandas MCP optimizes **One unified interface. 16 MCP servers. 150+ specialized tools. Built for research.** CLIO Kit is part of the IoWarp platform's comprehensive tooling ecosystem for AI agents. It brings AI assistance to your scientific computing workflow—whether you're analyzing terabytes of HDF5 data, managing Slurm jobs across clusters, or exploring research papers. Built by researchers, for researchers, at Illinois Institute of Technology with NSF support. > **Part of IoWarp Platform**: CLIO Kit is the tooling layer of the IoWarp platform, providing skills, plugins, and extensions for AI agents working in scientific computing environments. > **One simple command.** Production-ready, fully typed, MIT licensed, and beta-tested in real HPC environments. ## 🚀 Quick Installation ### One Command for Any Server ```bash # List all 16 available MCP servers uvx clio-kit mcp-servers # Run any server instantly uvx clio-kit mcp-server hdf5 uvx clio-kit mcp-server pandas uvx clio-kit mcp-server slurm # AI prompts also available uvx clio-kit prompts # List all prompts uvx clio-kit prompt code-coverage-prompt # Use a prompt ``` <details> <summary><b>Install in Cursor</b></summary> Add to your Cursor `~/.cursor/mcp.json`: ```json { "mcpServers": { "hdf5-mcp": { "command": "uvx", "args": ["clio-kit", "mcp-server", "hdf5"] }, "pandas-mcp": { "command": "uvx", "args": ["clio-kit", "mcp-server", "pandas"] }, "slurm-mcp": { "command": "uvx", "args": ["clio-kit", "mcp-server", "slurm"] } } } ``` See [Cursor MCP docs](https://docs.cursor.com/context/model-context-protocol) for more info. </details> <details> <summary><b>Install in Claude Code</b></summary> ```bash # Add HDF5 MCP claude mcp add hdf5-mcp -- uvx clio-kit mcp-server hdf5 # Add Pandas MCP claude mcp add pandas-mcp -- uvx clio-kit mcp-server pandas # Add Slurm MCP claude mcp add slurm-mcp -- uvx clio-kit mcp-server slurm ``` See [Claude Code MCP docs](https://docs.anthropic.com/en/docs/agents-and-tools/claude-code/tutorials#set-up-model-context-protocol-mcp) for more info. </details> <details> <summary><b>Install in VS Code</b></summary> Add to your VS Code MCP config: ```json "mcp": { "servers": { "hdf5-mcp": { "type": "stdio", "command": "uvx", "args": ["clio-kit", "mcp-server", "hdf5"] }, "pandas-mcp": { "type": "stdio", "command": "uvx", "args": ["clio-kit", "mcp-server", "pandas"] } } } ``` See [VS Code MCP docs](https://code.visualstudio.com/docs/copilot/chat/mcp-servers) for more info. </details> <details> <summary><b>Install in Claude Desktop</b></summary> Edit `claude_desktop_config.json`: ```json { "mcpServers": { "hdf5-mcp": { "command": "uvx", "args": ["clio-kit", "mcp-server", "hdf5"] }, "arxiv-mcp": { "command": "uvx", "args": ["clio-kit", "mcp-server", "arxiv"] } } } ``` See [Claude Desktop MCP docs](https://modelcontextprotocol.io/quickstart/user) for more info. </details> ## Available Packages <div align="center"> | 📦 **Package** | 📌 **Ver** | 🔧 **System** | 📋 **Description** | ⚡ **Install Command** | |:---|:---:|:---:|:---|:---| | **`adios`** | 1.0 | Data I/O | Read data using ADIOS2 engine | `uvx clio-kit mcp-server adios` | | **`arxiv`** | 1.0 | Research | Fetch research papers from ArXiv | `uvx clio-kit mcp-server arxiv` | | **`chronolog`** | 1.0 | Logging | Log and retrieve data from ChronoLog | `uvx clio-kit mcp-server chronolog` | | **`compression`** | 1.0 | Utilities | File compression with gzip | `uvx clio-kit mcp-server compression` | | **`darshan`** | 1.0 | Performance | I/O performance trace analysis | `uvx clio-kit mcp-server darshan` | | **`hdf5`** | 2.1 | Data I/O | HPC-optimized scientific data with 27 tools, AI insights, caching, streaming | `uvx clio-kit mcp-server hdf5` | | **`jarvis`** | 1.0 | Workflow | Data pipeline lifecycle management | `uvx clio-kit mcp-server jarvis` | | **`lmod`** | 1.0 | Environment | Environment module management | `uvx clio-kit mcp-server lmod` | | **`ndp`** | 1.0 | Data Protocol | Search and discover datasets across CKAN instances | `uvx clio-kit mcp-server ndp` | | **`node-hardware`** | 1.0 | System | System hardware information | `uvx clio-kit mcp-server node-hardware` | | **`pandas`** | 1.0 | Data Analysis | CSV data loading and filtering | `uvx clio-kit mcp-server pandas` | | **`parallel-sort`** | 1.0 | Computing | Large file sorting | `uvx clio-kit mcp-server parallel-sort` | | **`paraview`** | 1.0 | Visualization | Scientific 3D visualization and analysis | `uvx clio-kit mcp-server paraview` | | **`parquet`** | 1.0 | Data I/O | Read Parquet file columns | `uvx clio-kit mcp-server parquet` | | **`plot`** | 1.0 | Visualization | Generate plots from CSV data | `uvx clio-kit mcp-server plot` | | **`slurm`** | 1.0 | HPC | Job submission and management | `uvx clio-kit mcp-server slurm` | </div> --- ## 📖 Usage Examples ### HDF5: Scientific Data Analysis ``` "What datasets are in climate_simulation.h5? Show me the temperature field structure and read the first 100 timesteps." ``` **Tools used:** `open_file`, `analyze_dataset_structure`, `read_partial_dataset`, `list_attributes` ### Slurm: HPC Job Management ``` "Submit simulation.py to Slurm with 32 cores, 64GB memory, 24-hour runtime. Monitor progress and retrieve output when complete." ``` **Tools used:** `submit_slurm_job`, `check_job_status`, `get_job_output` ### ArXiv: Research Discovery ``` "Find the latest papers on diffusion models from ArXiv, get details on the top 3, and export citations to BibTeX." ``` **Tools used:** `search_arxiv`, `get_paper_details`, `export_to_bibtex`, `download_paper_pdf` ### Pandas: Data Processing ``` "Load sales_data.csv, clean missing values, compute statistics by region, and save as Parquet with compression." ``` **Tools used:** `load_data`, `handle_missing_data`, `groupby_operations`, `save_data` ### Plot: Data Visualization ``` "Create a line plot showing temperature trends over time from weather.csv with proper axis labels." ``` **Tools used:** `line_plot`, `data_info` --- ## 🚨 Troubleshooting <details> <summary><b>Server Not Found Error</b></summary> If `uvx clio-kit mcp-server <server-name>` fails: ```bash # Verify server name is correct uvx clio-kit mcp-servers # Common names: hdf5, pandas, slurm, arxiv (not hdf5-mcp, pandas-mcp) ``` </details> <details> <summary><b>Import Errors or Missing Dependencies</b></summary> For development or local testing: ```bash cd clio-kit-mcp-servers/hdf5 uv sync --all-extras --dev uv run hdf5-mcp ``` </details> <details> <summary><b>uvx Command Not Found</b></summary> Install uv package manager: ```bash # Linux/macOS curl -LsSf https://astral.sh/uv/install.sh | sh # Windows powershell -ExecutionPolicy ByPass -c "irm https://astral.sh/uv/install.ps1 | iex" # Or via pip pip install uv ``` </details> --- ## Team - **[Gnosis Research Center (GRC)](https://grc.iit.edu/)** - [Illinois Institute of Technology](https://www.iit.edu/) | Lead - **[HDF Group](https://www.hdfgroup.org/)** - Data format and library developers | Industry Partner - **[University of Utah](https://www.utah.edu/)** - Research collaboration | Domain Science Partner ## Sponsored By <img src="https://www.nsf.gov/themes/custom/nsf_theme/components/molecules/logo/logo-desktop.png" alt="NSF Logo" width="24" height="24"> **[NSF (National Science Foundation)](https://www.nsf.gov/)** - Supporting scientific computing research and AI integration initiatives > we welcome more sponsorships. please contact the [Principal Investigator](mailto:grc@illinoistech.edu) ## Ways to Contribute - **Submit Issues**: Report bugs or request features via [GitHub Issues](https://github.com/iowarp/clio-kit/issues) - **Develop New MCPs**: Add servers for your research tools ([CONTRIBUTING.md](CONTRIBUTING.md)) - **Improve Documentation**: Help make guides clearer - **Share Use Cases**: Tell us how you're using CLIO Kit in your research **Full Guide**: [CONTRIBUTING.md](CONTRIBUTING.md) ### Community & Support - **Chat**: [Zulip Community](https://iowarp.zulipchat.com/#narrow/channel/543872-Agent-Toolkit) - **Join**: [Invitation Link](https://iowarp.zulipchat.com/join/e4wh24du356e4y2iw6x6jeay/) - **Issues**: [GitHub Issues](https://github.com/iowarp/clio-kit/issues) - **Discussions**: [GitHub Discussions](https://github.com/iowarp/clio-kit/discussions) - **Website**: [https://toolkit.iowarp.ai/](https://toolkit.iowarp.ai/) - **Project**: [IOWarp Project](https://iowarp.ai) ---
text/markdown
null
IoWarp Team - Gnosis Research Center <grc@illinoistech.edu>
null
null
BSD-3-Clause
ai-agents, hdf5, hpc, llm-tools, mcp, model-context-protocol, parquet, research-tools, scientific-computing, slurm
[ "Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: Science/Research", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Topic :: Scientific/Engineering", "Topic :: Software Development :: Libraries :: Python Modules" ]
[]
null
null
>=3.10
[]
[]
[]
[ "click>=8.1.0" ]
[]
[]
[]
[ "Homepage, https://github.com/iowarp/clio-kit", "Repository, https://github.com/iowarp/clio-kit", "Issues, https://github.com/iowarp/clio-kit/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:28:43.546601
clio_kit-2.0.1.tar.gz
10,223,447
39/5c/50a141ac52ad6e0a4a630303a9834562ba71516e6586a4b75aa1fbc25cf7/clio_kit-2.0.1.tar.gz
source
sdist
null
false
718fd039d5d44f8292496e7822411e03
ea859e66e453c334d0b3371494089a99478257ce3a39e8d4ad16373217d6bb6e
395c50a141ac52ad6e0a4a630303a9834562ba71516e6586a4b75aa1fbc25cf7
null
[ "LICENSE" ]
174
2.3
anibridge-mal-provider
0.2.0b4
MAL provider for the AniBridge project.
# anibridge-mal-provider An [AniBridge](https://github.com/anibridge/anibridge) provider for [MyAnimeList](https://myanimelist.net/). _This provider comes built-in with AniBridge, so you don't need to install it separately._ ## Configuration ### `token` (`str`) Your MyAnimeList API refresh token. You can generate one [here](https://anibridge.eliasbenb.dev?generate_token=mal). ### `client_id` (`str`, optional) Your MyAnimeList API client ID. This option is for advanced users who want to use their own client ID. If not provided, a default client ID managed by the AniBridge team will be used. ```yaml list_provider_config: mal: token: ... client_id: "b11a4e1ead0db8142268906b4bb676a4" ```
text/markdown
Elias Benbourenane
Elias Benbourenane <eliasbenbourenane@gmail.com>
Elias Benbourenane
Elias Benbourenane <eliasbenbourenane@gmail.com>
null
null
[ "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.14", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent" ]
[]
null
null
>=3.14
[]
[]
[]
[ "aiohttp>=3.13.3", "anibridge-list-base>=0.2.0b3", "limiter>=0.5.0", "pydantic>=2.12.5" ]
[]
[]
[]
[]
uv/0.10.4 {"installer":{"name":"uv","version":"0.10.4","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
2026-02-21T00:28:19.947468
anibridge_mal_provider-0.2.0b4.tar.gz
8,589
ae/d4/16692fe64d6d4f71ccc02e855ab7b758de980fcfc638aca919ec74ca4f53/anibridge_mal_provider-0.2.0b4.tar.gz
source
sdist
null
false
799c56225fc82b807fb35a9675a507f3
94bc926e4cc7964edbf42aee47c3ab2ecef343e71d4432c8d89e92f1ec2e869d
aed416692fe64d6d4f71ccc02e855ab7b758de980fcfc638aca919ec74ca4f53
null
[]
152
2.1
pycarlo
0.12.158
Monte Carlo's Python SDK
# Pycarlo - Monte Carlo's Python SDK ## Installation Requires Python 3.9 or greater. Normally you can install and update using pip. For instance: ```shell virtualenv venv . venv/bin/activate pip install -U pycarlo ``` ## Overview Pycarlo comprises two components: `core` and `features`. All Monte Carlo API queries and mutations that you could execute via the API are supported via the `core` library. Operations can be executed as first class objects, using [sgqlc](https://github.com/profusion/sgqlc), or as raw GQL with variables. In both cases, a consistent object where fields can be referenced by dot notation and the more pythonic snake_case is returned for ease of use. The `features` library provides additional convenience for performing common operations like with dbt, circuit breaking, and pii filtering. Note that an API Key is required to use the SDK. See [our docs on generating API keys](https://docs.getmontecarlo.com/docs/developer-resources#creating-an-api-key) for details. ## Basic usage ### Core ```python from pycarlo.core import Client, Query, Mutation # First create a client. This creates a session using the 'default' profile from # '~/.mcd/profiles.ini'. This profile is created automatically via # `montecarlo configure` on the CLI. See the session subsection for # customizations, options and alternatives (e.g. using the environment, params, # named profiles, etc.) client = Client() # Now you can can execute a query. For instance, getUser (selecting the email field). # This would be like executing - # curl --location --request POST 'https://api.getmontecarlo.com/graphql' \ # --header 'x-mcd-id: <ID>' \ # --header 'x-mcd-token: <TOKEN>' \ # --header 'Content-Type: application/json' \ # --data-raw '{"query": "query {getUser {email}}"}' # Notice how the CamelCase from the Graphql query is converted to snake_case in # both the request and response. query = Query() query.get_user.__fields__('email') print(client(query).get_user.email) # You can also execute a query that requires variables. For instance, # testTelnetConnection (selecting all fields). query = Query() query.test_telnet_connection(host='montecarlodata.com', port=443) print(client(query)) # If necessary, you can always generate (e.g. print) the raw query that would be executed. print(query) # query { # testTelnetConnection(host: "montecarlodata.com", port: 443) { # success # validations { # type # message # } # warnings { # type # message # } # } # } # If you are not a fan of sgqlc operations (Query and Mutation) you can also execute any # raw query using the client. For instance, if we want the first 10 tables from getTables. get_table_query = """ query getTables{ getTables(first: 10) { edges { node { fullTableId } } } } """ response = client(get_table_query) # This returns a Box object where fields can be accessed using dot notation. # Notice how unlike with the API the response uses the more Pythonic snake_case. for edge in response.get_tables.edges: print(edge.node.full_table_id) # The response can still be processed as a standard dictionary. print(response['get_tables']['edges'][0]['node']['full_table_id']) # You can also execute any mutations too. For instance, generateCollectorTemplate # (selecting the templateLaunchUrl). mutation = Mutation() mutation.generate_collector_template().dc.template_launch_url() print(client(mutation)) # Any errors will raise a GqlError with details. For instance, executing above with an # invalid region. mutation = Mutation() mutation.generate_collector_template(region='artemis') print(client(mutation)) # pycarlo.common.errors.GqlError: [ # {'message': 'Region "\'artemis\'" not currently active.'... # ] ``` ### Examples #### Circuit Breaker Example ```python from pycarlo.core import Client, Session from pycarlo.features.circuit_breakers import CircuitBreakerService # Example from our test.snowflake account. endpoint = "https://api.dev.getmontecarlo.com/graphql" service = CircuitBreakerService( mc_client=Client(Session(mcd_profile="test-snow", endpoint=endpoint)), print_func=print ) in_breach = service.trigger_and_poll(rule_uuid="87872875-fe80-4963-8ab0-c04397a6daae") print("That can't be good. Our warehouse is broken." if in_breach else "Go, go, go!.") ``` #### Insight Upload Example ```python from pathlib import Path import boto3 import requests from pycarlo.core import Client, Query MC_CLIENT = Client() S3_CLIENT = boto3.client("s3") def upload_insights_to_s3( destination_bucket: str, desired_file_extension: str = ".csv", ) -> None: """ Example function for listing all insights in an account, and uploading any available to S3 as a CSV. """ list_insights_query = Query() list_insights_query.get_insights() for insight in MC_CLIENT(list_insights_query).get_insights: report_name = str(Path(insight.name).with_suffix(desired_file_extension)) if insight.available: report_url_query = Query() report_url_query.get_report_url(insight_name=insight.name, report_name=report_name) report_url = MC_CLIENT(report_url_query).get_report_url.url print(f"Uploading {report_name} to {destination_bucket}.") S3_CLIENT.upload_fileobj( Fileobj=requests.get(url=report_url, stream=True).raw, Bucket=destination_bucket, Key=report_name, ) if __name__ == "__main__": upload_insights_to_s3(destination_bucket="<BUCKET-NAME>") ``` See [Monte Carlo's API reference](https://apidocs.getmontecarlo.com/) for all supported queries and mutations. For details and additional examples on how to map (convert) GraphQL queries to `sgqlc` operations please refer to [the sgqlc docs](https://sgqlc.readthedocs.io/en/latest/sgqlc.operation.html). ### Features You can use [pydoc](https://docs.python.org/library/pydoc.html) to retrieve documentation on any feature packages (`pydoc pycarlo.features`). For instance for [circuit breakers](https://docs.getmontecarlo.com/docs/circuit-breakers): ```shell pydoc pycarlo.features.circuit_breakers.service ``` ## Session configuration By default, when creating a client the `default` profile from `~/.mcd/profiles.ini` is used. This file created via [montecarlo configure](https://docs.getmontecarlo.com/docs/using-the-cli#setting-up-the-cli) on the CLI. See [Monte Carlo's CLI reference](https://clidocs.getmontecarlo.com/) for more details. You can override this usage by creating a custom `Session`. For instance, if you want to pass the ID and Token: ```python from pycarlo.core import Client, Session client = Client(session=Session(mcd_id='foo', mcd_token='bar')) ``` Sessions support the following params: - `mcd_id`: API Key ID. - `mcd_token`: API secret. - `mcd_profile`: Named profile containing credentials. This is created via the CLI (e.g. `montecarlo configure --profile-name zeus`). - `mcd_config_path`: Path to file containing credentials. Defaults to `~/.mcd/`. You can also specify the API Key, secret or profile name using the following environment variables: - `MCD_DEFAULT_API_ID` - `MCD_DEFAULT_API_TOKEN` - `MCD_DEFAULT_PROFILE` When creating a session any explicitly passed `mcd_id` and `mcd_token` params take precedence, followed by environmental variables and then any config-file options. Environment variables can be mixed with passed credentials, but not the config-file profile. **We do not recommend passing `mcd_token` as it is a secret and can be accidentally committed.** ## Integration Gateway API There are features that require the Integration Gateway API instead of the regular GraphQL Application API, for example Airflow Callbacks invoked by the `airflow-mcd` library. To use the Gateway you need to initialize the `Session` object passing a `scope` parameter and then use `make_request` to invoke Gateway endpoints: ```python from pycarlo.core import Client, Session client = Client(session=Session(mcd_id='foo', mcd_token='bar', scope='AirflowCallbacks')) response = client.make_request( path='/airflow/callbacks', method='POST', body={}, timeout_in_seconds=20 ) ``` ## Advanced configuration The following values also be set by the environment: - `MCD_VERBOSE_ERRORS`: Enable logging. This includes a trace ID for each session and request. - `MCD_API_ENDPOINT`: Customize the endpoint where queries and mutations are executed. ## Enum Backward Compatibility Unlike the baseline `sgqlc` behavior, this SDK is designed to maintain backward compatibility when new enum values are added to the Monte Carlo API. If the API returns an enum value that doesn't exist in your SDK version, it will be returned as a string with a warning logged, rather than raising an error. This allows older SDK versions to continue working when new features are added. To avoid warnings and ensure full feature support, keep your SDK updated to the latest version. ## References - Monte Carlo App: <https://getmontecarlo.com> - Product docs: <https://docs.getmontecarlo.com> - Status page: <https://status.getmontecarlo.com> - API (and SDK): <https://apidocs.getmontecarlo.com> - CLI: <https://clidocs.getmontecarlo.com> ## License Apache 2.0 - See the [Apache License 2.0](https://www.apache.org/licenses/LICENSE-2.0) for more information.
text/markdown
Monte Carlo Data, Inc
info@montecarlodata.com
null
null
Apache Software License (Apache 2.0)
null
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "Natural Language :: English", "Topic :: Software Development :: Build Tools", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12" ]
[]
https://www.montecarlodata.com/
null
>=3.8
[]
[]
[]
[ "dataclasses_json<6.0.0,>=0.5.7", "python-box>=5.0.0", "requests<3.0.0,>=2.0.0", "responses>=0.20.0", "sgqlc<17.0,>=14.1" ]
[]
[]
[]
[]
twine/3.7.1 importlib_metadata/8.5.0 pkginfo/1.12.1.2 requests/2.32.4 requests-toolbelt/1.0.0 tqdm/4.67.3 CPython/3.8.6
2026-02-21T00:28:04.099188
pycarlo-0.12.158.tar.gz
1,205,251
42/1d/ebaa95fa7a97f0bf54a83e8ba36606218b41c35c2182893715daac04f9c4/pycarlo-0.12.158.tar.gz
source
sdist
null
false
6bb85ccbda3b60a8e8ec959b173dd4f6
577551eedccfbf210af8e3299567ed56f268692f1274e417b8c1cb1a0a98753e
421debaa95fa7a97f0bf54a83e8ba36606218b41c35c2182893715daac04f9c4
null
[]
9,702
2.3
anibridge-anilist-provider
0.2.0b5
AniList provider for the AniBridge project.
# anibridge-anilist-provider An [AniBridge](https://github.com/anibridge/anibridge) provider for [AniList](https://anilist.co/). _This provider comes built-in with AniBridge, so you don't need to install it separately._ ## Configuration ### `token` (`str`) Your AniList API token. You can generate one [here](https://anilist.co/login?apiVersion=v2&client_id=34003&response_type=token). ```yaml list_provider_config: anilist: token: ... ```
text/markdown
Elias Benbourenane
Elias Benbourenane <eliasbenbourenane@gmail.com>
Elias Benbourenane
Elias Benbourenane <eliasbenbourenane@gmail.com>
null
null
[ "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.14", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent" ]
[]
null
null
>=3.14
[]
[]
[]
[ "aiohttp>=3.13.2", "anibridge-list-base>=0.2.0b3", "async-lru>=2.0.5", "limiter>=0.5.0", "pydantic>=2.12.4" ]
[]
[]
[]
[]
uv/0.10.4 {"installer":{"name":"uv","version":"0.10.4","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
2026-02-21T00:28:02.011852
anibridge_anilist_provider-0.2.0b5-py3-none-any.whl
15,992
35/b3/7532405a09992ad9db058089494ae576477353f4f6982d33dd27ebf666d3/anibridge_anilist_provider-0.2.0b5-py3-none-any.whl
py3
bdist_wheel
null
false
37fb14d6c0f3d39f68d0fe2c8f2b7fe0
8ab42a0b266144eb327ac9baf44e8099464647ba4fb05a771bc6f59a98a89449
35b37532405a09992ad9db058089494ae576477353f4f6982d33dd27ebf666d3
null
[]
150
2.4
browseable
0.1.0
CLI for browser sessions and interactable element actions
# browseable `browseable` is a Unix-style CLI backed by a local daemon that keeps browser sessions alive between commands. ## What it does - API 1 (`elements`): given a session, list interactable form/navigation elements. - API 2 (`action`): given a session, perform an action (`click`, `type`, or `interact`). The daemon holds browser + page state so each CLI command is short-lived. ## Install ```bash uv sync uv run playwright install chromium ``` ## Quick start ```bash # Optional convenience. alias browseable='uv run browseable' # 1) start daemon browseable daemon start # 2) create a session (returns session ID) browseable session create https://example.com/signup # 3) list elements for that session browseable elements <session_id> # 4) focus an element browseable action <session_id> click 2 # 5) type into an element browseable action <session_id> type 3 "my@email.com" # 6) open interactive browser window for the same session browseable action <session_id> interact # 7) close session + stop daemon browseable session close <session_id> browseable daemon stop ``` ## Commands ```text browseable daemon start|run|status|stop browseable session create <url> browseable session close <session_id> browseable elements <session_id> [--json] browseable action <session_id> click <element_or_index> [--timeout-ms] browseable action <session_id> type <element_or_index> <text> [--timeout-ms] browseable action <session_id> interact [--timeout-ms] ``` `interact` focuses the session in a real browser window. If daemon started headless, `interact` promotes the session into a headed browser context. `BROWSEABLE_DAEMON_URL` can override the daemon address if needed. ## Daemon JSON API The daemon exposes local HTTP endpoints: - `POST /sessions` with `{"url":"..."}` -> create session. - `GET /sessions/<session_id>/elements` -> list interactable elements. - `POST /sessions/<session_id>/actions` with `{"action":"click"|"type","element_id":"...","text":"..."}`. - `POST /sessions/<session_id>/actions` with `{"action":"interact"}` -> promote/focus session in a headed browser window. - `DELETE /sessions/<session_id>` -> close session. ## E2E Example Run the telemetry register flow: ```bash uv run scripts/e2e_telemetry_register.sh ``` ## GitHub Workflows - `/.github/workflows/ci.yml`: compile + CLI smoke checks on `push`/`pull_request`. - `/.github/workflows/publish-testpypi.yml`: manual publish to TestPyPI. - `/.github/workflows/publish-pypi.yml`: publish to PyPI on `v*` tags (or manual dispatch). ## How to Publish to PyPI 1. Create package projects: - [https://test.pypi.org/](https://test.pypi.org/) - [https://pypi.org/](https://pypi.org/) 2. Configure Trusted Publishing on each project: - Owner/repo: your GitHub repo. - Workflow file: - TestPyPI: `.github/workflows/publish-testpypi.yml` - PyPI: `.github/workflows/publish-pypi.yml` - Environment name: - TestPyPI: `testpypi` - PyPI: `pypi` 3. Test release on TestPyPI by running `Publish to TestPyPI` (workflow dispatch). 4. Release to PyPI: - Bump version in `pyproject.toml`. - Commit and push. - Tag and push `vX.Y.Z` (must match `pyproject.toml` version). 5. Verify install: ```bash python -m pip install browseable==0.1.1 ``` ### Optional Manual Publish (No GitHub Actions) ```bash uv build uvx twine check dist/* uvx twine upload dist/* ``` For manual upload, create a PyPI API token and use `TWINE_USERNAME=__token__` and `TWINE_PASSWORD=<token>` (or `~/.pypirc`).
text/markdown
null
null
null
null
null
null
[]
[]
null
null
>=3.8
[]
[]
[]
[ "playwright<2,>=1.42.0" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:26:44.716571
browseable-0.1.0.tar.gz
13,073
d8/5a/6ee5a44edb67d5768a3284c2765e9ba95e8367b672fa5b5642a85b68ae52/browseable-0.1.0.tar.gz
source
sdist
null
false
708f1fd231fee99beb9bab51d2cad951
1e4e6398f22a885519c86e4a7d62abdbcce6bd2fc34f32884b0e56b199066e3e
d85a6ee5a44edb67d5768a3284c2765e9ba95e8367b672fa5b5642a85b68ae52
null
[]
179
2.4
trace-crispr
0.4.0
TRACE: Triple-aligner Read Analysis for CRISPR Editing
# TRACE **T**riple-aligner **R**ead **A**nalysis for **C**RISPR **E**diting TRACE is a comprehensive tool for quantifying CRISPR editing outcomes from amplicon sequencing data. It combines multiple alignment strategies with k-mer classification to provide robust, accurate measurements of HDR, NHEJ, and other editing outcomes. ## Features - **Triple-aligner consensus**: Uses BWA-MEM, BBMap, and minimap2 for robust alignment - **Flexible input**: Accepts DNA sequences directly or FASTA file paths - **Automatic inference**: Detects PAM, cleavage site, homology arms, and edits from sequences - **Large edit support**: Handles insertions up to 50+ bp with automatic k-mer size adjustment - **K-mer classification**: Fast pre-alignment HDR/WT detection (auto-sizes k-mers based on edit) - **Barcode-optimized k-mers**: Auto-detects barcode-style templates and generates discriminating k-mers - **Multi-nuclease support**: Cas9 and Cas12a (Cpf1) with correct cleavage geometry - **Robust UMI detection**: 3-pass algorithm handles variable primer quality and low-signal libraries - **Auto-detection**: Library type (TruSeq/Tn5), UMI presence, read merging need - **PCR deduplication**: Automatic UMI-based (TruSeq) or position-based (Tn5) deduplication - **CRISPResso2 integration**: Validation with standard CRISPR analysis tool ## Recent Updates ### Version 0.4.0 (2026-02-20) **New Features:** - **Alignment-only classification (now default)**: Pure alignment-based classification using multi-reference FASTA (WT + all HDR variants). More accurate than k-mer classification, especially for experiments with many similar barcodes. - **Full-length HDR builder**: Automatically builds full amplicon sequences from short donor templates using homology arm detection - **Global sequence deduplication**: Filter low-count sequences across all samples with `min_global_count` parameter - **Primary-only alignments**: When aligning to multi-reference FASTA, secondary alignments are suppressed by default **Breaking Changes:** - `alignment_only` parameter now defaults to `True`. Set `alignment_only=False` for legacy k-mer mode. **Bug Fixes:** - Fixed BBMap subprocess deadlock caused by stderr buffer overflow - Fixed version mismatch between `__init__.py` and `pyproject.toml` See [CHANGELOG.md](CHANGELOG.md) for full details. ### Version 0.3.1 (2026-02-17) **Critical Bug Fixes:** - **3-pass UMI detection algorithm**: Dramatically improves merge rates for libraries with weak primer signals - Pass 1: Strong signal detection (>50% consensus) - high confidence - Pass 2: Weak signal detection (>30% consensus, ≥4bp UMI) - medium confidence - Pass 3: Jump detection with 6bp fallback - handles poor quality data - **Impact**: Merge rate improved from 35% to 92.5% on HEK293 EMX1 test dataset (80 samples) **Optimizations:** - **Barcode-style template auto-detection**: Automatically identifies when templates share homology arms with different barcodes - **Optimized k-mer generation**: Generates k-mers that span barcode boundaries for better discrimination - **Multi-template batch processing**: ~100x faster classification through global sequence deduplication **Dependencies:** - Added CRISPResso2 to optional validation dependencies (`pip install trace-crispr[validation]`) ## Installation ### pip (Python package only) ```bash pip install trace-crispr ``` ### conda (includes external aligners) ```bash conda install -c bioconda -c conda-forge trace-crispr ``` ### Development installation ```bash git clone https://github.com/k-roy/TRACE.git cd TRACE pip install -e ".[dev]" ``` ## Quick Start TRACE accepts sequences as either **DNA strings** or **FASTA file paths**. ### Example 1: Using FASTA files ```bash trace run \ --reference amplicon.fasta \ --hdr-template hdr_template.fasta \ --guide GCTGAAGCACTGCACGCCGT \ --r1 sample_R1.fastq.gz \ --r2 sample_R2.fastq.gz \ --output results/ ``` ### Example 2: Using DNA sequences directly The reference amplicon will typically be longer than the HDR template so that the primers specifically amplify the target locus. This example shows a 250 bp reference sequence where the donor template is 150 bp long with the designed edit in the middle. ```bash # Reference amplicon (250 bp) - includes flanking regions # Guide sequence shown in lowercase for illustration REF="ATCGATCGATCGATCGATCGATCGATCGATCGATCGATCGATCGATCGATCG\ ATCGATCGATCGATCGATCGATCGATCGATCGATCGATCGATCGATCGATCG\ gctgaagcactgcacgccgttgg\ ATCGATCGATCGATCGATCGATCGATCGATCGATCGATCGATCGATCGATCG\ ATCGATCGATCGATCGATCGATCGATCGATCGATCGATCGATCGATCGATCG" # HDR template (150 bp) - centered on edit site # Guide in lowercase, designed edit (T->A) shown in UPPERCASE HDR="ATCGATCGATCGATCGATCGATCGATCGATCGATCGATCGATCGATCGATCG\ gctgaagcactgcacgccgtAga\ ATCGATCGATCGATCGATCGATCGATCGATCGATCGATCGATCGATCGATCG" trace run \ -r "$REF" \ -h "$HDR" \ -g GCTGAAGCACTGCACGCCGT \ --r1 sample_R1.fastq.gz \ --output results/ ``` **Note:** The guide and edit are shown here in lowercase/uppercase for illustrative purposes. This formatting is not necessary - TRACE will automatically detect guide and edit positions from the sequences. ### Check locus configuration without running ```bash trace info \ --reference amplicon.fasta \ --hdr-template hdr_template.fasta \ --guide GCTGAAGCACTGCACGCCGT ``` This will print: ``` ============================================================ === TRACE Analysis Configuration === ============================================================ Reference sequence: 231 bp HDR template: 127 bp - Template aligns at position 53 in reference Donor template analysis: - Left homology arm: positions 53-124 on reference (72 bp) - Right homology arm: positions 128-179 on reference (52 bp) Edits detected (2 total): * Position 125: T -> A (substitution) * Position 127: G -> A (substitution) Guide analysis: - Guide sequence: GCTGAAGCACTGCACGCCGT - Guide targets: positions 105-124 on reference (+ strand) - PAM: TGG at positions 125-127 on reference - Cleavage site: position 122 on reference ``` ### Multiple samples Create a sample key TSV: ``` sample_id r1_path r2_path condition sample_1 /path/to/S1_R1.fastq.gz /path/to/S1_R2.fastq.gz treatment sample_2 /path/to/S2_R1.fastq.gz /path/to/S2_R2.fastq.gz control ``` Then run: ```bash trace run \ --reference amplicon.fasta \ --hdr-template hdr_template.fasta \ --guide GCTGAAGCACTGCACGCCGT \ --sample-key samples.tsv \ --output results/ \ --threads 16 ``` ### Per-sample locus sequences For multiplexed experiments with different target loci, you can specify `reference`, `hdr_template`, and `guide` per-sample in the sample key TSV: ``` sample_id r1_path r2_path condition reference hdr_template guide sample_1 /path/S1_R1.fq.gz /path/S1_R2.fq.gz treatment sample_2 /path/S2_R1.fq.gz /path/S2_R2.fq.gz control sample_3 /path/S3_R1.fq.gz /path/S3_R2.fq.gz treatment locus2.fasta locus2_hdr.fasta ACGTACGTACGTACGTACGT sample_4 /path/S4_R1.fq.gz /path/S4_R2.fq.gz control locus2.fasta locus2_hdr.fasta ACGTACGTACGTACGTACGT ``` - **Empty values**: Use CLI defaults (`--reference`, `--hdr-template`, `--guide`) - **Filled values**: Override CLI defaults for that sample - Values can be DNA sequences or FASTA file paths ### Multi-Template Analysis (Barcode Screening) For barcode screening experiments where multiple HDR templates (barcodes) are possible, use the `multi-template` command: ```bash # Generate HDR templates FASTA from keyfiles trace generate-templates \ --sample-key keyfiles/sample_key.tsv \ --seq-ref keyfiles/guide_donor_and_reference_info.tsv \ --output templates/hdr_templates.fasta # Generate sample manifest from keyfiles trace generate-manifest \ --sample-key keyfiles/sample_key.tsv \ --plate-key keyfiles/plate_key.tsv \ --raw-data-dir raw_data/ \ --output trace_sample_key.tsv # Run multi-template analysis trace multi-template \ --reference templates/reference.fasta \ --hdr-templates templates/hdr_templates.fasta \ --guide GAGTCCGAGCAGAAGAAGAA \ --sample-key trace_sample_key.tsv \ --output results/ \ --threads 16 ``` **Output tables:** - `per_sample_editing_outcomes_all_methods.tsv` - Summary per sample - `per_sample_per_template_outcomes.tsv` - Granular per-template results **Features:** - Detects which barcode/template is present in each read - Purity checking (detects unexpected barcodes) - AMBIGUOUS category for reads matching multiple barcodes - Expected template validation via `expected_barcode` column - Parallel processing with configurable threads ### Using Cas12a ```bash trace run \ --reference amplicon.fasta \ --hdr-template hdr_template.fasta \ --guide GCTGAAGCACTGCACGCCGTAA \ --nuclease cas12a \ --sample-key samples.tsv \ --output results/ ``` ## Auto-Detection TRACE automatically detects library characteristics to optimize analysis: ### Library Type Detection TRACE distinguishes between **TruSeq** (fixed-target amplicon) and **Tn5** (tagmented locus) libraries by analyzing read alignment positions: - **TruSeq**: Reads cluster at fixed start positions (primer binding sites) - **Tn5**: Reads have scattered start positions (random Tn5 cutting) ``` Auto-detection results: - Library type: TruSeq (100% of reads cluster at fixed start position) - UMI detection: UMIs of length 6 bp detected --> Entering PCR deduplication mode... ``` ### UMI Detection For TruSeq libraries, TRACE detects UMIs (Unique Molecular Identifiers) by analyzing sequence diversity at read starts: - High diversity region = UMI - Low diversity region = primer sequence - Automatically determines UMI length (typically 4-12 bp) ### Preprocessing Modes Based on detection results, TRACE automatically selects the optimal preprocessing workflow: | Library | UMIs | Overlap (>=15bp) | Preprocessing | Output | |---------|------|------------------|---------------|--------| | TruSeq | Yes | Yes | dedup → trim → merge → collapse | merged FASTQ | | TruSeq | Yes | No | dedup → trim | paired FASTQs | | TruSeq | No | Yes | trim → merge → collapse | merged FASTQ | | TruSeq | No | No | trim | paired FASTQs | | Tn5 | N/A | Yes | trim → merge → collapse → align → position dedup | merged FASTQ | | Tn5 | N/A | No | trim → align → position dedup | paired FASTQs | Example output: ``` Auto-detection results: - Library type: TruSeq (100% of reads cluster at fixed start position) - UMI detection: UMIs of length 6 bp detected - Read overlap: Enabled (~50bp overlap (25% of amplicon)) - Preprocessing: dedup-trim-merge-collapse -> merged (UMI dedup -> trim -> merge -> collapse) - CRISPResso mode: merged (merged reads from preprocessing) ``` ## Designed Edit Detection TRACE automatically detects the edits encoded in the donor by first aligning the HDR template to the reference. TRACE then classifies the intended edit as a single-nucleotide variant (SNV), multi-nucleotide variant (MNV), insertion, or deletion. K-mers are selected that span the designed edit and are unique to the reference and donor. For large edits, TRACE automatically increases the k-mer size to ensure reliable classification. TRACE can handle MNVs or insertions with lengths up to the read length - 50 bp (e.g., 100 bp insertions can be detected with 150 bp reads). Example output for a 20 bp insertion: ``` Edits detected (1 total): * Position 125: +ATCGATCGATCGATCGATCG (20 bp insertion) Maximum edit size: 20 bp Recommended k-mer size: 30 bp ``` ## Nuclease Support ### Cas9 (SpCas9) - PAM: NGG (3' of protospacer) - Cleavage: 3 bp upstream of PAM (blunt ends) ### Cas12a (LbCpf1) - PAM: TTTN (5' of protospacer) - Cleavage: 18-19 bp downstream on target strand, 23 bp on non-target - Creates 4-5 nt 5' overhang (staggered cut) ## Output The main output is a TSV file with per-sample editing outcomes: | Column | Description | |--------|-------------| | sample | Sample ID | | classifiable_reads | Total classifiable reads | | duplicate_rate | PCR duplicate rate | | WT_% | Wild-type % | | HDR_% | HDR % | | NHEJ_% | NHEJ % | | LgDel_% | Large deletion % | | kmer_hdr_rate | K-mer method HDR rate | | crispresso_hdr_rate | CRISPResso2 HDR rate | | crispresso_indel_rate | CRISPResso2 indel rate | For Tn5/tagmented data or TruSeq amplicons with UMIs, TRACE will report on the PCR duplication rate and automatically perform deduplication: - **TruSeq with UMIs**: Pre-alignment UMI-based deduplication - **Tn5**: Post-alignment position-based deduplication ## Analysis and Visualization TRACE includes an analysis module for comparing editing outcomes across conditions with statistical testing and publication-quality visualizations. ### Installation The analysis module requires additional dependencies for visualization: ```bash pip install trace-crispr[visualization] ``` Or install scipy, matplotlib, and seaborn separately: ```bash pip install scipy matplotlib seaborn ``` ### Basic Usage #### Compare conditions from pipeline results ```python from trace_crispr.analysis import ( compare_metric_by_condition, results_to_dataframe, get_condition_stats, plot_condition_comparison, ) # After running the TRACE pipeline results = pipeline.run_all(samples) # Compare HDR rates across conditions comparisons = compare_metric_by_condition( results, samples, condition_col='treatment', # Column in sample metadata metric='dedup_hdr_pct', # Metric to compare base_condition='control' # Reference condition for t-tests ) # View results as a DataFrame print(comparisons.to_dataframe()) ``` Output: ``` condition base_condition metric condition_mean condition_std ... p_value p_adjusted significance 0 treatment_A control dedup_hdr_pct 25.41 2.63 ... 0.0003 0.0006 *** 1 treatment_B control dedup_hdr_pct 12.15 1.89 ... 0.4521 0.4521 ns ``` #### Create bar plots with replicate points ```python # Convert results to DataFrame and get stats df = results_to_dataframe(results, samples) stats = get_condition_stats(df, 'treatment', 'dedup_hdr_pct') # Create bar plot with individual points and significance stars fig = plot_condition_comparison( stats, comparisons, base_condition='control', title='HDR Rate by Treatment', ylabel='HDR Rate (%)' ) fig.savefig('hdr_comparison.png', dpi=150, bbox_inches='tight') ``` This creates a bar chart showing: - Mean values as bars - Individual replicate values as overlaid points (with jitter) - Standard error of mean (SEM) as error bars - Significance stars above significantly different conditions #### Work directly with DataFrames If you already have a DataFrame (e.g., from a previous analysis): ```python from trace_crispr.analysis import ( compare_dataframe_by_condition, get_condition_stats, plot_condition_comparison, ) import pandas as pd # Load existing data df = pd.read_csv('editing_outcomes.tsv', sep='\t') # Compare conditions comparisons = compare_dataframe_by_condition( df, condition_col='treatment', metric='dedup_hdr_pct', base_condition='control' ) # Get stats and plot stats = get_condition_stats(df, 'treatment', 'dedup_hdr_pct') fig = plot_condition_comparison(stats, comparisons) ``` #### Get summary statistics ```python from trace_crispr.analysis import get_condition_summary # Generate summary table for all metrics summary = get_condition_summary(results, samples, condition_col='treatment') print(summary[['condition', 'n', 'dedup_hdr_pct_mean', 'dedup_hdr_pct_sem']]) ``` Output: ``` condition n dedup_hdr_pct_mean dedup_hdr_pct_sem 0 control 4 10.25 0.68 1 treatment_A 4 25.41 1.32 2 treatment_B 4 12.15 0.94 ``` ### Statistical Methods - **T-test**: Welch's t-test (unequal variances) comparing each condition to the base - **FDR correction**: Benjamini-Hochberg correction applied by default (disable with `fdr_correction=False`) - **Significance thresholds**: `*` (p < 0.05), `**` (p < 0.01), `***` (p < 0.001) ### Available Functions | Function | Description | |----------|-------------| | `compare_metric_by_condition()` | Main entry point - compare a metric across conditions from SampleResults | | `compare_dataframe_by_condition()` | Compare conditions from an existing DataFrame | | `get_condition_summary()` | Generate summary statistics table | | `results_to_dataframe()` | Convert SampleResult list to DataFrame | | `get_condition_stats()` | Calculate mean, std, sem, n for each condition | | `compare_conditions()` | Perform statistical comparisons between conditions | | `plot_condition_comparison()` | Bar plot with points, error bars, and significance stars | | `plot_comparison_summary()` | Forest/bar plot of fold changes | | `plot_replicate_correlation()` | Scatter plot comparing two metrics | | `plot_multi_metric_comparison()` | Multi-panel comparison across metrics | ### Plot Customization ```python fig = plot_condition_comparison( stats, comparisons, base_condition='control', title='HDR Rate by Treatment', ylabel='HDR Rate (%)', figsize=(12, 6), # Figure size bar_color='#22c55e', # Color for treatment bars (green) base_color='#888888', # Color for base condition (gray) point_alpha=0.6, # Transparency for data points jitter=0.15, # Horizontal spread for points show_significance=True, # Show significance stars condition_order=['control', 'treatment_A', 'treatment_B'], # Custom order ) ``` ## Dependencies ### Python (Core) - click>=8.0 - pysam>=0.20 - pandas>=1.5 - numpy>=1.20 - pyyaml>=6.0 - rapidfuzz>=3.0 - tqdm>=4.60 ### Python (Visualization - optional) - matplotlib>=3.5 - seaborn>=0.12 - scipy>=1.9 Install with: `pip install trace-crispr[visualization]` ### External tools (via conda) - bwa>=0.7 - bbmap>=39 - minimap2>=2.24 - samtools>=1.16 - crispresso2 (optional, but enabled by default) ## Author Kevin R. Roy ## License MIT
text/markdown
null
"Kevin R. Roy" <kevinroy@stanford.edu>
null
null
null
null
[ "Development Status :: 4 - Beta", "Intended Audience :: Science/Research", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Topic :: Scientific/Engineering :: Bio-Informatics" ]
[]
null
null
>=3.9
[]
[]
[]
[ "click>=8.0", "pysam>=0.20", "pandas>=1.5", "numpy>=1.20", "pyyaml>=6.0", "rapidfuzz>=3.0", "tqdm>=4.60", "pytest>=7.0; extra == \"dev\"", "pytest-cov; extra == \"dev\"", "black; extra == \"dev\"", "ruff; extra == \"dev\"", "mypy; extra == \"dev\"", "matplotlib>=3.5; extra == \"visualization\"", "seaborn>=0.12; extra == \"visualization\"", "scipy>=1.9; extra == \"visualization\"", "CRISPResso2>=2.2; extra == \"validation\"" ]
[]
[]
[]
[ "Homepage, https://github.com/k-roy/trace", "Documentation, https://trace-crispr.readthedocs.io", "Repository, https://github.com/k-roy/trace" ]
twine/6.2.0 CPython/3.12.2
2026-02-21T00:25:22.974431
trace_crispr-0.4.0.tar.gz
107,884
93/d0/7d1602c20bc83ab73e2d1d4f993ee7e13933288b7e0dde7de1a8c3a9892a/trace_crispr-0.4.0.tar.gz
source
sdist
null
false
a7353f0bd31de4b1a910cf692e518ebf
5ca1a53495f54eee50e05691020d344aaca6be40901f731a21f916c1f6329b8f
93d07d1602c20bc83ab73e2d1d4f993ee7e13933288b7e0dde7de1a8c3a9892a
MIT
[ "LICENSE" ]
169
2.4
pytorch-ignite
0.6.0.dev20260221
A lightweight library to help with training neural networks in PyTorch.
<div align="center"> <!-- ![Ignite Logo](assets/logo/ignite_logo_mixed.svg) --> <img src="https://raw.githubusercontent.com/pytorch/ignite/master/assets/logo/ignite_logo_mixed.svg" width=512> <!-- [![image](https://travis-ci.com/pytorch/ignite.svg?branch=master)](https://travis-ci.com/pytorch/ignite) --> | ![image](https://img.shields.io/badge/-Tests:-black?style=flat-square) [![image](https://github.com/pytorch/ignite/actions/workflows/unit-tests.yml/badge.svg?branch=master)](https://github.com/pytorch/ignite/actions/workflows/unit-tests.yml) [![image](https://github.com/pytorch/ignite/actions/workflows/gpu-tests.yml/badge.svg)](https://github.com/pytorch/ignite/actions/workflows/gpu-tests.yml) [![image](https://codecov.io/gh/pytorch/ignite/branch/master/graph/badge.svg)](https://codecov.io/gh/pytorch/ignite) [![image](https://img.shields.io/badge/dynamic/json.svg?label=docs&url=https%3A%2F%2Fpypi.org%2Fpypi%2Fpytorch-ignite%2Fjson&query=%24.info.version&colorB=brightgreen&prefix=v)](https://pytorch.org/ignite/index.html) | |:--- | ![image](https://img.shields.io/badge/-Stable%20Releases:-black?style=flat-square) [![image](https://anaconda.org/pytorch/ignite/badges/version.svg)](https://anaconda.org/pytorch/ignite) ・ [![image](https://img.shields.io/badge/dynamic/json.svg?label=PyPI&url=https%3A%2F%2Fpypi.org%2Fpypi%2Fpytorch-ignite%2Fjson&query=%24.info.version&colorB=brightgreen&prefix=v)](https://pypi.org/project/pytorch-ignite/) [![image](https://static.pepy.tech/badge/pytorch-ignite)](https://pepy.tech/project/pytorch-ignite) ・ [![image](https://img.shields.io/badge/docker-hub-blue)](https://hub.docker.com/u/pytorchignite) | | ![image](https://img.shields.io/badge/-Nightly%20Releases:-black?style=flat-square) [![image](https://anaconda.org/pytorch-nightly/ignite/badges/version.svg)](https://anaconda.org/pytorch-nightly/ignite) [![image](https://img.shields.io/badge/PyPI-pre%20releases-brightgreen)](https://pypi.org/project/pytorch-ignite/#history)| | ![image](https://img.shields.io/badge/-Community:-black?style=flat-square) [![Twitter](https://img.shields.io/badge/news-twitter-blue)](https://twitter.com/pytorch_ignite) [![discord](https://img.shields.io/badge/chat-discord-blue?logo=discord)](https://discord.gg/djZtm3EmKj) [![numfocus](https://img.shields.io/badge/NumFOCUS-affiliated%20project-green)](https://numfocus.org/sponsored-projects/affiliated-projects) | | ![image](https://img.shields.io/badge/-Supported_PyTorch/Python_versions:-black?style=flat-square) [![link](https://img.shields.io/badge/-check_here-blue)](https://github.com/pytorch/ignite/actions?query=workflow%3A%22PyTorch+version+tests%22)| </div> ## TL;DR Ignite is a high-level library to help with training and evaluating neural networks in PyTorch flexibly and transparently. <div align="center"> <a href="https://colab.research.google.com/github/pytorch/ignite/blob/master/assets/tldr/teaser.ipynb"> <img alt="PyTorch-Ignite teaser" src="https://raw.githubusercontent.com/pytorch/ignite/master/assets/tldr/pytorch-ignite-teaser.gif" width=532> </a> _Click on the image to see complete code_ </div> ### Features - [Less code than pure PyTorch](https://raw.githubusercontent.com/pytorch/ignite/master/assets/ignite_vs_bare_pytorch.png) while ensuring maximum control and simplicity - Library approach and no program's control inversion - _Use ignite where and when you need_ - Extensible API for metrics, experiment managers, and other components <!-- ############################################################################################################### --> # Table of Contents - [Table of Contents](#table-of-contents) - [Why Ignite?](#why-ignite) - [Simplified training and validation loop](#simplified-training-and-validation-loop) - [Power of Events & Handlers](#power-of-events--handlers) - [Execute any number of functions whenever you wish](#execute-any-number-of-functions-whenever-you-wish) - [Built-in events filtering](#built-in-events-filtering) - [Stack events to share some actions](#stack-events-to-share-some-actions) - [Custom events to go beyond standard events](#custom-events-to-go-beyond-standard-events) - [Out-of-the-box metrics](#out-of-the-box-metrics) - [Installation](#installation) - [Nightly releases](#nightly-releases) - [Docker Images](#docker-images) - [Using pre-built images](#using-pre-built-images) - [Getting Started](#getting-started) - [Documentation](#documentation) - [Additional Materials](#additional-materials) - [Examples](#examples) - [Tutorials](#tutorials) - [Reproducible Training Examples](#reproducible-training-examples) - [Communication](#communication) - [User feedback](#user-feedback) - [Contributing](#contributing) - [Projects using Ignite](#projects-using-ignite) - [Citing Ignite](#citing-ignite) - [About the team & Disclaimer](#about-the-team--disclaimer) <!-- ############################################################################################################### --> # Why Ignite? Ignite is a **library** that provides three high-level features: - Extremely simple engine and event system - Out-of-the-box metrics to easily evaluate models - Built-in handlers to compose training pipeline, save artifacts and log parameters and metrics ## Simplified training and validation loop No more coding `for/while` loops on epochs and iterations. Users instantiate engines and run them. <details> <summary> Example </summary> ```python from ignite.engine import Engine, Events, create_supervised_evaluator from ignite.metrics import Accuracy # Setup training engine: def train_step(engine, batch): # Users can do whatever they need on a single iteration # Eg. forward/backward pass for any number of models, optimizers, etc # ... trainer = Engine(train_step) # Setup single model evaluation engine evaluator = create_supervised_evaluator(model, metrics={"accuracy": Accuracy()}) def validation(): state = evaluator.run(validation_data_loader) # print computed metrics print(trainer.state.epoch, state.metrics) # Run model's validation at the end of each epoch trainer.add_event_handler(Events.EPOCH_COMPLETED, validation) # Start the training trainer.run(training_data_loader, max_epochs=100) ``` </details> ## Power of Events & Handlers The cool thing with handlers is that they offer unparalleled flexibility (compared to, for example, callbacks). Handlers can be any function: e.g. lambda, simple function, class method, etc. Thus, we do not require to inherit from an interface and override its abstract methods which could unnecessarily bulk up your code and its complexity. ### Execute any number of functions whenever you wish <details> <summary> Examples </summary> ```python trainer.add_event_handler(Events.STARTED, lambda _: print("Start training")) # attach handler with args, kwargs mydata = [1, 2, 3, 4] logger = ... def on_training_ended(data): print(f"Training is ended. mydata={data}") # User can use variables from another scope logger.info("Training is ended") trainer.add_event_handler(Events.COMPLETED, on_training_ended, mydata) # call any number of functions on a single event trainer.add_event_handler(Events.COMPLETED, lambda engine: print(engine.state.times)) @trainer.on(Events.ITERATION_COMPLETED) def log_something(engine): print(engine.state.output) ``` </details> ### Built-in events filtering <details> <summary> Examples </summary> ```python # run the validation every 5 epochs @trainer.on(Events.EPOCH_COMPLETED(every=5)) def run_validation(): # run validation # change some training variable once on 20th epoch @trainer.on(Events.EPOCH_STARTED(once=20)) def change_training_variable(): # ... # Trigger handler with customly defined frequency @trainer.on(Events.ITERATION_COMPLETED(event_filter=first_x_iters)) def log_gradients(): # ... ``` </details> ### Stack events to share some actions <details> <summary> Examples </summary> Events can be stacked together to enable multiple calls: ```python @trainer.on(Events.COMPLETED | Events.EPOCH_COMPLETED(every=10)) def run_validation(): # ... ``` </details> ### Custom events to go beyond standard events <details> <summary> Examples </summary> Custom events related to backward and optimizer step calls: ```python from ignite.engine import EventEnum class BackpropEvents(EventEnum): BACKWARD_STARTED = 'backward_started' BACKWARD_COMPLETED = 'backward_completed' OPTIM_STEP_COMPLETED = 'optim_step_completed' def update(engine, batch): # ... loss = criterion(y_pred, y) engine.fire_event(BackpropEvents.BACKWARD_STARTED) loss.backward() engine.fire_event(BackpropEvents.BACKWARD_COMPLETED) optimizer.step() engine.fire_event(BackpropEvents.OPTIM_STEP_COMPLETED) # ... trainer = Engine(update) trainer.register_events(*BackpropEvents) @trainer.on(BackpropEvents.BACKWARD_STARTED) def function_before_backprop(engine): # ... ``` - Complete snippet is found [here](https://pytorch.org/ignite/faq.html#creating-custom-events-based-on-forward-backward-pass). - Another use-case of custom events: [trainer for Truncated Backprop Through Time](https://pytorch.org/ignite/contrib/engines.html#ignite.contrib.engines.create_supervised_tbptt_trainer). </details> ## Out-of-the-box metrics - [Metrics](https://pytorch.org/ignite/metrics.html#complete-list-of-metrics) for various tasks: Precision, Recall, Accuracy, Confusion Matrix, IoU etc, ~20 [regression metrics](https://pytorch.org/ignite/metrics.html#complete-list-of-metrics). - Users can also [compose their metrics](https://pytorch.org/ignite/metrics.html#metric-arithmetics) with ease from existing ones using arithmetic operations or torch methods. <details> <summary> Example </summary> ```python precision = Precision(average=False) recall = Recall(average=False) F1_per_class = (precision * recall * 2 / (precision + recall)) F1_mean = F1_per_class.mean() # torch mean method F1_mean.attach(engine, "F1") ``` </details> <!-- ############################################################################################################### --> # Installation From [pip](https://pypi.org/project/pytorch-ignite/): ```bash pip install pytorch-ignite ``` From [conda](https://anaconda.org/pytorch/ignite): ```bash conda install ignite -c pytorch ``` From source: ```bash pip install git+https://github.com/pytorch/ignite ``` ## Nightly releases From pip: ```bash pip install --pre pytorch-ignite ``` From conda (this suggests to install [pytorch nightly release](https://anaconda.org/pytorch-nightly/pytorch) instead of stable version as dependency): ```bash conda install ignite -c pytorch-nightly ``` ## Docker Images ### Using pre-built images Pull a pre-built docker image from [our Docker Hub](https://hub.docker.com/u/pytorchignite) and run it with docker v19.03+. ```bash docker run --gpus all -it -v $PWD:/workspace/project --network=host --shm-size 16G pytorchignite/base:latest /bin/bash ``` <details> <summary> List of available pre-built images </summary> Base - `pytorchignite/base:latest` - `pytorchignite/apex:latest` - `pytorchignite/hvd-base:latest` - `pytorchignite/hvd-apex:latest` - `pytorchignite/msdp-apex:latest` Vision: - `pytorchignite/vision:latest` - `pytorchignite/hvd-vision:latest` - `pytorchignite/apex-vision:latest` - `pytorchignite/hvd-apex-vision:latest` - `pytorchignite/msdp-apex-vision:latest` NLP: - `pytorchignite/nlp:latest` - `pytorchignite/hvd-nlp:latest` - `pytorchignite/apex-nlp:latest` - `pytorchignite/hvd-apex-nlp:latest` - `pytorchignite/msdp-apex-nlp:latest` </details> For more details, see [here](docker). <!-- ############################################################################################################### --> # Getting Started Few pointers to get you started: - [Quick Start Guide: Essentials of getting a project up and running](https://pytorch-ignite.ai/tutorials/beginner/01-getting-started/) - [Concepts of the library: Engine, Events & Handlers, State, Metrics](https://pytorch-ignite.ai/concepts/) - Full-featured template examples (coming soon) <!-- ############################################################################################################### --> # Documentation - Stable API documentation and an overview of the library: https://pytorch.org/ignite/ - Development version API documentation: https://pytorch.org/ignite/master/ - [FAQ](https://pytorch.org/ignite/faq.html), ["Questions on Github"](https://github.com/pytorch/ignite/issues?q=is%3Aissue+label%3Aquestion+) and ["Questions on Discuss.PyTorch"](https://discuss.pytorch.org/c/ignite). - [Project's Roadmap](https://github.com/pytorch/ignite/wiki/Roadmap) ## Additional Materials - [Distributed Training Made Easy with PyTorch-Ignite](https://labs.quansight.org/blog/2021/06/distributed-made-easy-with-ignite/) - [PyTorch Ecosystem Day 2021 Breakout session presentation](https://colab.research.google.com/drive/1qhUgWQ0N2U71IVShLpocyeY4AhlDCPRd) - [Tutorial blog post about PyTorch-Ignite](https://labs.quansight.org/blog/2020/09/pytorch-ignite/) - [8 Creators and Core Contributors Talk About Their Model Training Libraries From PyTorch Ecosystem](https://neptune.ai/blog/model-training-libraries-pytorch-ecosystem?utm_source=reddit&utm_medium=post&utm_campaign=blog-model-training-libraries-pytorch-ecosystem) - Ignite Posters from Pytorch Developer Conferences: - [2021](https://drive.google.com/file/d/1YXrkJIepPk_KltSG1ZfWRtA5IRgPFz_U) - [2019](https://drive.google.com/open?id=1bqIl-EM6GCCCoSixFZxhIbuF25F2qTZg) - [2018](https://drive.google.com/open?id=1_2vzBJ0KeCjGv1srojMHiJRvceSVbVR5) <!-- ############################################################################################################### --> # Examples ## Tutorials - [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/pytorch/ignite/blob/master/examples/notebooks/TextCNN.ipynb) [Text Classification using Convolutional Neural Networks](https://github.com/pytorch/ignite/blob/master/examples/notebooks/TextCNN.ipynb) - [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/pytorch/ignite/blob/master/examples/notebooks/VAE.ipynb) [Variational Auto Encoders](https://github.com/pytorch/ignite/blob/master/examples/notebooks/VAE.ipynb) - [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/pytorch/ignite/blob/master/examples/notebooks/FashionMNIST.ipynb) [Convolutional Neural Networks for Classifying Fashion-MNIST Dataset](https://github.com/pytorch/ignite/blob/master/examples/notebooks/FashionMNIST.ipynb) - [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/pytorch/ignite/blob/master/examples/notebooks/CycleGAN_with_nvidia_apex.ipynb) [Training Cycle-GAN on Horses to Zebras with Nvidia/Apex](https://github.com/pytorch/ignite/blob/master/examples/notebooks/CycleGAN_with_nvidia_apex.ipynb) - [ logs on W&B](https://app.wandb.ai/vfdev-5/ignite-cyclegan-apex) - [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/pytorch/ignite/blob/master/examples/notebooks/CycleGAN_with_torch_cuda_amp.ipynb) [Another training Cycle-GAN on Horses to Zebras with Native Torch CUDA AMP](https://github.com/pytorch/ignite/blob/master/examples/notebooks/CycleGAN_with_torch_cuda_amp.ipynb) - [logs on W&B](https://app.wandb.ai/vfdev-5/ignite-cyclegan-torch-amp) - [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/pytorch/ignite/blob/master/examples/notebooks/EfficientNet_Cifar100_finetuning.ipynb) [Finetuning EfficientNet-B0 on CIFAR100](https://github.com/pytorch/ignite/blob/master/examples/notebooks/EfficientNet_Cifar100_finetuning.ipynb) - [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/pytorch/ignite/blob/master/examples/notebooks/Cifar10_Ax_hyperparam_tuning.ipynb) [Hyperparameters tuning with Ax](https://github.com/pytorch/ignite/blob/master/examples/notebooks/Cifar10_Ax_hyperparam_tuning.ipynb) - [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/pytorch/ignite/blob/master/examples/notebooks/FastaiLRFinder_MNIST.ipynb) [Basic example of LR finder on MNIST](https://github.com/pytorch/ignite/blob/master/examples/notebooks/FastaiLRFinder_MNIST.ipynb) - [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/pytorch/ignite/blob/master/examples/notebooks/Cifar100_bench_amp.ipynb) [Benchmark mixed precision training on Cifar100: torch.amp vs nvidia/apex](https://github.com/pytorch/ignite/blob/master/examples/notebooks/Cifar100_bench_amp.ipynb) - [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/pytorch/ignite/blob/master/examples/notebooks/MNIST_on_TPU.ipynb) [MNIST training on a single TPU](https://github.com/pytorch/ignite/blob/master/examples/notebooks/MNIST_on_TPU.ipynb) - [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1E9zJrptnLJ_PKhmaP5Vhb6DTVRvyrKHx) [CIFAR10 Training on multiple TPUs](https://github.com/pytorch/ignite/tree/master/examples/cifar10) - [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/pytorch/ignite/blob/master/examples/notebooks/HandlersTimeProfiler_MNIST.ipynb) [Basic example of handlers time profiling on MNIST training example](https://github.com/pytorch/ignite/blob/master/examples/notebooks/HandlersTimeProfiler_MNIST.ipynb) ## Reproducible Training Examples Inspired by [torchvision/references](https://github.com/pytorch/vision/tree/master/references), we provide several reproducible baselines for vision tasks: - [ImageNet](examples/references/classification/imagenet) - logs on Ignite Trains server coming soon ... - [Pascal VOC2012](examples/references/segmentation/pascal_voc2012) - logs on Ignite Trains server coming soon ... Features: - Distributed training: native or horovod and using [PyTorch native AMP](https://pytorch.org/docs/stable/notes/amp_examples.html) ## Code-Generator application The easiest way to create your training scripts with PyTorch-Ignite: - https://code-generator.pytorch-ignite.ai/ <!-- ############################################################################################################### --> # Communication - [GitHub issues](https://github.com/pytorch/ignite/issues): questions, bug reports, feature requests, etc. - [Discuss.PyTorch](https://discuss.pytorch.org/c/ignite), category "Ignite". - [PyTorch-Ignite Discord Server](https://discord.gg/djZtm3EmKj): to chat with the community - [GitHub Discussions](https://github.com/pytorch/ignite/discussions): general library-related discussions, ideas, Q&A, etc. ## User feedback We have created a form for ["user feedback"](https://github.com/pytorch/ignite/issues/new/choose). We appreciate any type of feedback, and this is how we would like to see our community: - If you like the project and want to say thanks, this the right place. - If you do not like something, please, share it with us, and we can see how to improve it. Thank you! <!-- ############################################################################################################### --> # Contributing Please see the [contribution guidelines](https://github.com/pytorch/ignite/blob/master/CONTRIBUTING.md) for more information. As always, PRs are welcome :) <!-- ############################################################################################################### --> # Projects using Ignite <details> <summary> Research papers </summary> - [BatchBALD: Efficient and Diverse Batch Acquisition for Deep Bayesian Active Learning](https://github.com/BlackHC/BatchBALD) - [A Model to Search for Synthesizable Molecules](https://github.com/john-bradshaw/molecule-chef) - [Localised Generative Flows](https://github.com/jrmcornish/lgf) - [Extracting T Cell Function and Differentiation Characteristics from the Biomedical Literature](https://github.com/hammerlab/t-cell-relation-extraction) - [Variational Information Distillation for Knowledge Transfer](https://github.com/amzn/xfer/tree/master/var_info_distil) - [XPersona: Evaluating Multilingual Personalized Chatbot](https://github.com/HLTCHKUST/Xpersona) - [CNN-CASS: CNN for Classification of Coronary Artery Stenosis Score in MPR Images](https://github.com/ucuapps/CoronaryArteryStenosisScoreClassification) - [Bridging Text and Video: A Universal Multimodal Transformer for Video-Audio Scene-Aware Dialog](https://github.com/ictnlp/DSTC8-AVSD) - [Adversarial Decomposition of Text Representation](https://github.com/text-machine-lab/adversarial_decomposition) - [Uncertainty Estimation Using a Single Deep Deterministic Neural Network](https://github.com/y0ast/deterministic-uncertainty-quantification) - [DeepSphere: a graph-based spherical CNN](https://github.com/deepsphere/deepsphere-pytorch) - [Norm-in-Norm Loss with Faster Convergence and Better Performance for Image Quality Assessment](https://github.com/lidq92/LinearityIQA) - [Unified Quality Assessment of In-the-Wild Videos with Mixed Datasets Training](https://github.com/lidq92/MDTVSFA) - [Deep Signature Transforms](https://github.com/patrick-kidger/Deep-Signature-Transforms) - [Neural CDEs for Long Time-Series via the Log-ODE Method](https://github.com/jambo6/neuralCDEs-via-logODEs) - [Volumetric Grasping Network](https://github.com/ethz-asl/vgn) - [Mood Classification using Listening Data](https://github.com/fdlm/listening-moods) - [Deterministic Uncertainty Estimation (DUE)](https://github.com/y0ast/DUE) - [PyTorch-Hebbian: facilitating local learning in a deep learning framework](https://github.com/Joxis/pytorch-hebbian) - [Stochastic Weight Matrix-Based Regularization Methods for Deep Neural Networks](https://github.com/rpatrik96/lod-wmm-2019) - [Learning explanations that are hard to vary](https://github.com/gibipara92/learning-explanations-hard-to-vary) - [The role of disentanglement in generalisation](https://github.com/mmrl/disent-and-gen) - [A Probabilistic Programming Approach to Protein Structure Superposition](https://github.com/LysSanzMoreta/Theseus-PP) - [PadChest: A large chest x-ray image dataset with multi-label annotated reports](https://github.com/auriml/Rx-thorax-automatic-captioning) </details> <details> <summary> Blog articles, tutorials, books </summary> - [State-of-the-Art Conversational AI with Transfer Learning](https://github.com/huggingface/transfer-learning-conv-ai) - [Tutorial on Transfer Learning in NLP held at NAACL 2019](https://github.com/huggingface/naacl_transfer_learning_tutorial) - [Deep-Reinforcement-Learning-Hands-On-Second-Edition, published by Packt](https://github.com/PacktPublishing/Deep-Reinforcement-Learning-Hands-On-Second-Edition) - [Once Upon a Repository: How to Write Readable, Maintainable Code with PyTorch](https://towardsdatascience.com/once-upon-a-repository-how-to-write-readable-maintainable-code-with-pytorch-951f03f6a829) - [The Hero Rises: Build Your Own SSD](https://allegro.ai/blog/the-hero-rises-build-your-own-ssd/) - [Using Optuna to Optimize PyTorch Ignite Hyperparameters](https://medium.com/pytorch/using-optuna-to-optimize-pytorch-ignite-hyperparameters-626ffe6d4783) - [PyTorch Ignite - Classifying Tiny ImageNet with EfficientNet](https://towardsdatascience.com/pytorch-ignite-classifying-tiny-imagenet-with-efficientnet-e5b1768e5e8f) </details> <details> <summary> Toolkits </summary> - [Project MONAI - AI Toolkit for Healthcare Imaging](https://github.com/Project-MONAI/MONAI) - [DeepSeismic - Deep Learning for Seismic Imaging and Interpretation](https://github.com/microsoft/seismic-deeplearning) - [Nussl - a flexible, object-oriented Python audio source separation library](https://github.com/nussl/nussl) - [PyTorch Adapt - A fully featured and modular domain adaptation library](https://github.com/KevinMusgrave/pytorch-adapt) - [gnina-torch: PyTorch implementation of GNINA scoring function](https://github.com/RMeli/gnina-torch) </details> <details> <summary> Others </summary> - [Implementation of "Attention is All You Need" paper](https://github.com/akurniawan/pytorch-transformer) - [Implementation of DropBlock: A regularization method for convolutional networks in PyTorch](https://github.com/miguelvr/dropblock) - [Kaggle Kuzushiji Recognition: 2nd place solution](https://github.com/lopuhin/kaggle-kuzushiji-2019) - [Unsupervised Data Augmentation experiments in PyTorch](https://github.com/vfdev-5/UDA-pytorch) - [Hyperparameters tuning with Optuna](https://github.com/optuna/optuna-examples/blob/main/pytorch/pytorch_ignite_simple.py) - [Logging with ChainerUI](https://chainerui.readthedocs.io/en/latest/reference/module.html#external-library-support) - [FixMatch experiments in PyTorch and Ignite (CTA dataaug policy)](https://github.com/vfdev-5/FixMatch-pytorch) - [Kaggle Birdcall Identification Competition: 1st place solution](https://github.com/ryanwongsa/kaggle-birdsong-recognition) - [Logging with Aim - An open-source experiment tracker](https://aimstack.readthedocs.io/en/latest/quick_start/integrations.html#integration-with-pytorch-ignite) </details> See other projects at ["Used by"](https://github.com/pytorch/ignite/network/dependents?package_id=UGFja2FnZS02NzI5ODEwNA%3D%3D) If your project implements a paper, represents other use-cases not covered in our official tutorials, Kaggle competition's code, or just your code presents interesting results and uses Ignite. We would like to add your project to this list, so please send a PR with brief description of the project. <!-- ############################################################################################################### --> # Citing Ignite If you use PyTorch-Ignite in a scientific publication, we would appreciate citations to our project. ``` @misc{pytorch-ignite, author = {V. Fomin and J. Anmol and S. Desroziers and J. Kriss and A. Tejani}, title = {High-level library to help with training neural networks in PyTorch}, year = {2020}, publisher = {GitHub}, journal = {GitHub repository}, howpublished = {\url{https://github.com/pytorch/ignite}}, } ``` <!-- ############################################################################################################### --> # About the team & Disclaimer PyTorch-Ignite is a [NumFOCUS Affiliated Project](https://www.numfocus.org/), operated and maintained by volunteers in the PyTorch community in their capacities as individuals (and not as representatives of their employers). See the ["About us"](https://pytorch-ignite.ai/about/community/#about-us) page for a list of core contributors. For usage questions and issues, please see the various channels [here](#communication). For all other questions and inquiries, please send an email to contact@pytorch-ignite.ai.
text/markdown
null
PyTorch-Ignite Team <contact@pytorch-ignite.ai>
null
null
null
null
[ "Programming Language :: Python :: 3" ]
[]
null
null
<=3.14,>=3.10
[]
[]
[]
[ "packaging", "torch<3,>=1.10" ]
[]
[]
[]
[ "Homepage, https://pytorch-ignite.ai", "Repository, https://github.com/pytorch/ignite" ]
twine/6.2.0 CPython/3.10.19
2026-02-21T00:24:59.569945
pytorch_ignite-0.6.0.dev20260221.tar.gz
7,509,379
fa/3c/fc718f68b9a2ee11fc6e6671574fd7b806f256218c32d0c0f0b9a7decb1a/pytorch_ignite-0.6.0.dev20260221.tar.gz
source
sdist
null
false
77a2e323ce7280e556491b864d1dd55b
d40fb43f68035e03abe191075e8d38618ea40ee1d51a13ca367e68b32c1a93ab
fa3cfc718f68b9a2ee11fc6e6671574fd7b806f256218c32d0c0f0b9a7decb1a
BSD-3-Clause
[ "LICENSE" ]
198
2.4
evolve-sdk
0.0.26
Pythonic SDK for multi-agent orchestration in E2B sandboxes
# Evolve SDK Python SDK Evolve SDK lets you run and orchestrate terminal-based AI agents in secure sandboxes with built-in observability. Check out the [official documentation](https://github.com/evolving-machines-lab/evolve/tree/main/docs) and [cookbooks](https://github.com/evolving-machines-lab/evolve/tree/main/cookbooks). ## Reporting Bugs We welcome your feedback. File a [GitHub issue](https://github.com/evolving-machines-lab/evolve/issues) to report bugs or request features. ## Connect on Discord Join the [Evolve SDK Developers Discord](https://discord.gg/Q36D8dGyNF) to connect with other developers using Evolve SDK. Get help, share feedback, and discuss your projects with the community. ## License See [LICENSE](https://github.com/evolving-machines-lab/evolve/blob/main/LICENSE) for details.
text/markdown
null
"Swarmlink, Inc." <brandomagnani@evolvingmachines.ai>
null
null
Apache-2.0
ai, agents, sandbox, e2b, orchestration, codex, claude, gemini, evolve-sdk
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Topic :: Software Development :: Libraries :: Python Modules" ]
[]
null
null
>=3.10
[]
[]
[]
[ "pytest>=8.0.0; extra == \"dev\"", "pytest-asyncio>=0.23.0; extra == \"dev\"", "python-dotenv>=1.0.0; extra == \"dev\"", "pydantic>=2.0.0; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://github.com/evolving-machines-lab/evolve", "Repository, https://github.com/evolving-machines-lab/evolve" ]
twine/6.2.0 CPython/3.11.14
2026-02-21T00:22:28.234244
evolve_sdk-0.0.26.tar.gz
1,058,225
21/1e/1c946d20d7c4d9d26f4c242a850e68a19fee94a35fc45e97e51c1c1b3a16/evolve_sdk-0.0.26.tar.gz
source
sdist
null
false
ca720892078783fd1ea5b0f15c0b3fdd
3bea6c2b211c238067b9b40513fe5e82174e0420da04ba9d770134c939ae1863
211e1c946d20d7c4d9d26f4c242a850e68a19fee94a35fc45e97e51c1c1b3a16
null
[ "LICENSE" ]
192
2.4
uipath
2.8.49
Python SDK and CLI for UiPath Platform, enabling programmatic interaction with automation services, process management, and deployment tools.
# UiPath Python SDK [![PyPI downloads](https://img.shields.io/pypi/dm/uipath.svg)](https://pypi.org/project/uipath/) [![PyPI - Version](https://img.shields.io/pypi/v/uipath)](https://img.shields.io/pypi/v/uipath) [![Python versions](https://img.shields.io/pypi/pyversions/uipath.svg)](https://pypi.org/project/uipath/) A Python SDK that enables programmatic interaction with UiPath Cloud Platform services including processes, assets, buckets, context grounding, data services, jobs, and more. The package also features a CLI for creation, packaging, and deployment of automations to UiPath Cloud Platform. Use the [UiPath LangChain SDK](https://github.com/UiPath/uipath-langchain-python) to pack and publish LangGraph Agents. Check out other [UiPath Integrations](https://github.com/UiPath/uipath-integrations-python) to pack and publish agents built with LlamaIndex, OpenAI Agents, Google ADK, and more. This [quickstart guide](https://uipath.github.io/uipath-python/) walks you through deploying your first agent to UiPath Cloud Platform. ## Table of Contents - [Installation](#installation) - [Configuration](#configuration) - [Environment Variables](#environment-variables) - [Basic Usage](#basic-usage) - [Available Services](#available-services) - [Examples](#examples) - [Buckets Service](#buckets-service) - [Context Grounding Service](#context-grounding-service) - [Command Line Interface (CLI)](#command-line-interface-cli) - [Authentication](#authentication) - [Initialize a Project](#initialize-a-project) - [Debug a Project](#debug-a-project) - [Package a Project](#package-a-project) - [Publish a Package](#publish-a-package) - [Project Structure](#project-structure) - [Development](#development) - [Setting Up a Development Environment](#setting-up-a-development-environment) ## Installation ```bash pip install uipath ``` using `uv`: ```bash uv add uipath ``` ## Configuration ### Environment Variables Create a `.env` file in your project root with the following variables: ``` UIPATH_URL=https://cloud.uipath.com/ACCOUNT_NAME/TENANT_NAME UIPATH_ACCESS_TOKEN=YOUR_TOKEN_HERE ``` ## Basic Usage ```python from uipath.platform import UiPath # Initialize the SDK sdk = UiPath() # Execute a process job = sdk.processes.invoke( name="MyProcess", input_arguments={"param1": "value1", "param2": 42} ) # Work with assets asset = sdk.assets.retrieve(name="MyAsset") ``` ## Available Services The SDK provides access to various UiPath services: - `sdk.processes` - Manage and execute UiPath automation processes - `sdk.assets` - Work with assets (variables, credentials) stored in UiPath - `sdk.buckets` - Manage cloud storage containers for automation files - `sdk.connections` - Handle connections to external systems - `sdk.context_grounding` - Work with semantic contexts for AI-enabled automation - `sdk.jobs` - Monitor and manage automation jobs - `sdk.queues` - Work with transaction queues - `sdk.tasks` - Work with Action Center - `sdk.api_client` - Direct access to the API client for custom requests ## Examples ### Buckets Service ```python # Download a file from a bucket sdk.buckets.download( bucket_key="my-bucket", blob_file_path="path/to/file.xlsx", destination_path="local/path/file.xlsx" ) ``` ### Context Grounding Service ```python # Search for contextual information results = sdk.context_grounding.search( name="my-knowledge-index", query="How do I process an invoice?", number_of_results=5 ) ``` ## Command Line Interface (CLI) The SDK also provides a command-line interface for creating, packaging, and deploying automations: ### Authentication ```bash uipath auth ``` This command opens a browser for authentication and creates/updates your `.env` file with the proper credentials. ### Initialize a Project ```bash uipath init ``` The `uipath.json` file should include your entry points in the `functions` section: ```json { "functions": { "main": "main.py:main" } } ``` Running `uipath init` will process these function definitions and create the corresponding `entry-points.json` file needed for deployment. For more details on the configuration format, see the [UiPath configuration specifications](specs/README.md). ### Debug a Project ```bash uipath run ENTRYPOINT [INPUT] ``` Executes a Python script with the provided JSON input arguments. ### Package a Project ```bash uipath pack ``` Packages your project into a `.nupkg` file that can be deployed to UiPath. **Note:** Your `pyproject.toml` must include: - A description field (avoid characters: &, <, >, ", ', ;) - Author information Example: ```toml description = "Your package description" authors = [{name = "Your Name", email = "your.email@example.com"}] ``` ### Publish a Package ```bash uipath publish ``` Publishes the most recently created package to your UiPath Orchestrator. ## Project Structure To properly use the CLI for packaging and publishing, your project should include: - A `pyproject.toml` file with project metadata - A `uipath.json` file with your function definitions (e.g., `"functions": {"main": "main.py:main"}`) - A `entry-points.json` file (generated by `uipath init`) - A `bindings.json` file (generated by `uipath init`) to configure resource overrides - Any Python files needed for your automation ## Development ### Tools Check out [uipath-dev](https://github.com/uipath/uipath-dev-python) - an interactive application for building, testing, and debugging UiPath Python runtimes, agents, and automation scripts. ### Contributions Please read our [contribution guidelines](https://github.com/UiPath/uipath-integrations-python/blob/main/CONTRIBUTING.md) before submitting a pull request.
text/markdown
null
null
null
Marius Cosareanu <marius.cosareanu@uipath.com>, Cristian Pufu <cristian.pufu@uipath.com>
null
null
[ "Intended Audience :: Developers", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Software Development :: Build Tools" ]
[]
null
null
>=3.11
[]
[]
[]
[ "applicationinsights>=0.11.10", "click>=8.3.1", "coverage>=7.8.2", "graphtty==0.1.8", "httpx>=0.28.1", "mermaid-builder==0.0.3", "mockito>=1.5.4", "pathlib>=1.0.1", "pydantic-function-models>=0.1.11", "pyjwt>=2.10.1", "pysignalr==1.3.0", "python-dotenv>=1.0.1", "python-socketio<6.0.0,>=5.15.0", "rich>=14.2.0", "tenacity>=9.0.0", "truststore>=0.10.1", "uipath-core<0.6.0,>=0.5.0", "uipath-runtime<0.10.0,>=0.9.0" ]
[]
[]
[]
[ "Homepage, https://uipath.com", "Repository, https://github.com/UiPath/uipath-python", "Documentation, https://uipath.github.io/uipath-python/" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:22:18.185101
uipath-2.8.49.tar.gz
4,607,263
c3/0e/f1d41d28df1c2a54a8ce72fa52ab7ccce17377b50f4a10545244a6cda04c/uipath-2.8.49.tar.gz
source
sdist
null
false
d4b66ce7d4fa7e624535f1c0b428eac4
3fd49cc4e6cb2d337a59a5866a0eab8459d3006bc8df27fafeeadc742ae7bc44
c30ef1d41d28df1c2a54a8ce72fa52ab7ccce17377b50f4a10545244a6cda04c
null
[ "LICENSE" ]
7,810
2.4
SimpleGeometryCalc
0.0.6.2
A small library to calculate most variables of shapes from just one!
# SimpleGeometry A simple python library to calculate variables of shapes with only one input. # Commands formulas(shape) -- formulas allows you to rewiew the formulas of either a square or a circle (if you forgot them) allesumcircle(r=,d=,c=,a=) -- allesumcircle allows you to input either the radius (r) or the diameter (d) or the circumference (c) or the area (a) and gain access to all the other variables with the simple command. you can also turn printing off by doing Print=False. allesumsquare(s=,a=,p=,d=, Print=False) -- allesumsquare allows you to input either the sides (s) or the area (a) or the perimeter (p) or the diagonal (d) and gain access to all the other variables. you can also turn printing off by doing Print=False. Variables that come with this library: pi -- the pi variable is just your standard pi (3.14159265358979323846264338) # Pip now available via pip install SimpleGeometryCalc https://pypi.org/project/SimpleGeometryCalc/#description 👍
text/markdown
Tatpoato
null
Tatpoato
null
null
Geometry, Simple, calc, geometry, simple
[]
[]
null
null
null
[]
[]
[]
[]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:22:05.495275
simplegeometrycalc-0.0.6.2.tar.gz
2,973
ba/bd/8b4e5b3ede789275a98cd260ef249c3587e2a63186c6d547a6bdad5a59da/simplegeometrycalc-0.0.6.2.tar.gz
source
sdist
null
false
6678900d74608dcb1ccbffc1399cebd1
2a1483bc6753ea7669265a3b22f1562c5aaec27a811f2bee2acc258ed0ba8329
babd8b4e5b3ede789275a98cd260ef249c3587e2a63186c6d547a6bdad5a59da
null
[ "LICENSE" ]
0
2.4
hedgehog
1.0.13
Hierarchical Evaluation of Drug GEnerators tHrOugh riGorous filtration
# 🦔 HEDGEHOG **Hierarchical Evaluation of Drug GEnerators tHrOugh riGorous filtration** [![PyPI version](https://badge.fury.io/py/hedgehog.svg)](https://pypi.org/project/hedgehog/) [![CI](https://github.com/LigandPro/hedgehog/actions/workflows/ci.yaml/badge.svg)](https://github.com/LigandPro/hedgehog/actions/workflows/ci.yaml) [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/licenses/MIT) [![Python 3.10+](https://img.shields.io/badge/python-3.10+-blue.svg)](https://www.python.org/downloads/) ![HEDGEHOG Pipeline](data/imgs/pipeline_structure.png) Comprehensive benchmark pipeline for evaluating generative models in molecular design. ### Pipeline Stages: Each stage takes the output of the previous one, progressively filtering the molecule set: 1) **Mol Prep (Datamol)**: salts/solvents & fragments cleanup, largest-fragment selection, metal disconnection, uncharging, tautomer canonicalization, stereochemistry removal → produces standardized “clean” molecules ([molPrep folder](src/hedgehog/stages/molPrep/)) 2) **Molecular Descriptors**: 22 physicochemical descriptors (logP, HBD/HBA, TPSA, QED, etc.) → molecules outside thresholds are removed ([descriptors folder](src/hedgehog/stages/descriptors/)) 3) **Structural Filters**: 6 criteria with ~2500 SMARTS patterns (PAINS, Glaxo, NIBR, Bredt, etc.) → flagged molecules are removed ([structural filters folder](src/hedgehog/stages/structFilters/)) 4) **Synthesis Evaluation**: SA score, SYBA score, AiZynthFinder retrosynthesis → unsynthesizable molecules are removed ([synthesis folder](src/hedgehog/stages/synthesis/)) 5) **Molecular Docking**: SMINA and/or GNINA → binding affinity scoring ([docking folder](src/hedgehog/stages/docking/)) 6) **Docking Filters**: post-docking pose quality filtering → poor binders are removed 7) **Final Descriptors**: recalculation on the filtered set Post-pipeline analysis: MolEval generative metrics ## Setup & Run ### Install from PyPI ```bash python -m pip install hedgehog hedgehog --help ``` Base install is intentionally lightweight and works on modern Python versions (including Python 3.13) without optional heavy docking extras. Optional extras: ```bash # Legacy PoseCheck backend for docking filters python -m pip install 'hedgehog[docking-legacy]' # Shepherd-Score Python dependency only (may be unavailable on some ABIs, e.g. cp313) python -m pip install 'hedgehog[shepherd]' ``` Recommended Shepherd setup is an isolated worker environment: ```bash uv run hedgehog setup shepherd-worker --yes ``` ### Install from source (recommended for development) ```bash # Clone repository git clone https://github.com/LigandPro/hedgehog.git cd hedgehog # Install AiZynthFinder (for synthesis stage) - recommended CLI flow uv run hedgehog setup aizynthfinder # Legacy helper script (alternative) ./modules/install_aizynthfinder.sh # Install package with uv uv sync ``` You are ready to use **🦔 HEDGEHOG** for your purpose! **Usage** ```bash # Run full pipeline on a proposed small test data from `data/test/` uv run hedgehog run # Alternatively, using the short-name alias: uv run hedge run # Run full pipeline on your own molecule file uv run hedge run --mols data/my_molecules.csv # Run full pipeline on your own molecule files via glob uv run hedge run --mols "data/generated/*.csv" # Run specific stage uv run hedge run --stage descriptors # Run a specific stage on your own molecule file uv run hedge run --stage descriptors --mols data/my_molecules.csv # Auto-install missing optional external tools during a run uv run hedge run --auto-install # Reuse the existing results folder uv run hedge run --reuse # Force a fresh results folder for stage reruns uv run hedge run --stage docking --force-new # Enable live progress bar in CLI uv run hedge run --progress # Regenerate HTML report from an existing run uv run hedge report results/run_10 # Show pipeline stages and current version uv run hedge info uv run hedge version # Get help uv run hedge --help ``` **GNINA (CPU/GPU) notes** GNINA is auto-resolved during the docking stage: - If `gnina` is already on `PATH`, HEDGEHOG uses it. - Otherwise it auto-downloads a compatible Linux GNINA binary to `~/.hedgehog/bin/gnina`. By default, auto-install prefers the CPU variant. To prefer CUDA builds: ```bash export HEDGEHOG_GNINA_VARIANT=auto # or: gpu uv run hedge run --stage docking --auto-install ``` GNINA runtime now auto-discovers CUDA/PyTorch libraries from common locations (including `site-packages/nvidia/*/lib`, active conda env, and `~/miniforge/lib`) so manual `gnina_ld_library_path` is usually not required. **Terminal UI (TUI)** For interactive configuration and pipeline management, use the TUI: ```bash uv run hedgehog tui ``` If the TUI has not been built yet, the CLI will install/build it automatically on first launch. You can also launch it directly from the TUI package: ```bash cd tui npm run tui ``` See [tui/README.md](tui/README.md) for details and developer workflow. **Unified verification pipeline** Use one command entry point for local/CI checks: ```bash # Quick local smoke (CLI + TUI build + TUI startup/quit in PTY) uv run python scripts/check_pipeline.py --mode quick # CI smoke profile (same checks, no full production run) uv run python scripts/check_pipeline.py --mode ci # Full local verification (quick checks + full production pipeline run) uv run python scripts/check_pipeline.py --mode full ``` `--mode full` runs `uv run hedgehog run` with the default production config, so it can be long-running and requires external stage dependencies (for example docking/synthesis tooling) to be installed in your local environment. **Git hooks with Lefthook (recommended)** Use Lefthook to block commits/pushes that would fail CI: ```bash # Install Lefthook (macOS) brew install lefthook # Register git hooks from lefthook.yml lefthook install # Optional: run hooks manually lefthook run pre-commit lefthook run pre-push ``` Current local gates: - `pre-commit`: staged Python formatting/lint (`ruff`) and whitespace checks. - `pre-push`: repository-wide `ruff` checks, `pytest`, pipeline smoke (`scripts/check_pipeline.py --mode ci`), and docs build (`docs && pnpm build`). If you need to skip a specific hook command once (not recommended), use `SKIP`: ```bash SKIP=docs-build git push ``` **Documentation Site** ```bash cd docs && pnpm install && pnpm dev ``` The docs site is built with [Nextra](https://nextra.site) and available at `http://localhost:3000`. **HTML Reports** After each pipeline run, an interactive HTML report is automatically generated as `report.html` in the results folder. The report includes: - Pipeline summary and molecule retention funnel - Per-stage statistics and visualizations - Descriptor distributions - Filter pass/fail breakdowns - Synthesis scores and docking results **Configure your run** Edit config for each stage in [configs folder](src/hedgehog/configs/) based on metrics you want to calculate. <!-- ## REINVENT4 fine-tune To fine-tune REINVENT4 follow these steps: 1. Clone REINVENT4 repository and setup environment: ```bash git clone https://github.com/MolecularAI/REINVENT4.git ``` 2. Configure transfer learning (aka fine-tuning) 1. Adjust `configs/toml/transfer_learning.toml` following provided `configs/toml/README.md` instructions, 2. Set input model file for Mol2Mol generator as provided by authors `priors/reinvent.prior`. 3. Set the following parameters: ```ini num_epochs = 1000 save_every_n_epochs = 10 batch_size = [adjust appropriately to reduce training time] ``` 3) Train the model: Run the training using the modified configuration file. It takes approximetely 72 hours to train a model on ~750 samples with that setup. Once trained, fine-tuned model can be used for downstream evaluation and benchmarking tasks. --- -->
text/markdown
null
null
null
null
null
null
[]
[]
null
null
>=3.10
[]
[]
[]
[ "pandas>=2.0.0", "numpy>=1.21.0", "matplotlib>=3.4.0", "seaborn>=0.11.0", "pyyaml>=5.4.0", "rdkit>=2022.3.0", "datamol>=0.8.0", "medchem>=0.1.0", "rich>=13.0.0", "typer>=0.9.0", "syba", "jinja2>=3.0.0", "plotly>=5.0.0", "prolif>=2.1.0", "xgboost>=2", "posecheck-fast>=0.1.6; python_full_version >= \"3.11\"", "ruff; extra == \"lint\"", "mypy; extra == \"lint\"", "pyright; extra == \"lint\"", "pytest; extra == \"test\"", "pytest-cov; extra == \"test\"", "shepherd-score>=1.2.1; extra == \"shepherd\"", "posecheck>=1.3.1; extra == \"docking-legacy\"" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:21:49.071353
hedgehog-1.0.13.tar.gz
11,530,705
2f/7b/368e75217a1f088bae9eab832a75a6914ab02bcf513fa1ed9d76e811a3cc/hedgehog-1.0.13.tar.gz
source
sdist
null
false
d0acdbb41229e2f4feb835f2667d01cd
9d17233b8b5065eb4d653ff45eb6657be06cfb65ffd5679059654f0797c9a707
2f7b368e75217a1f088bae9eab832a75a6914ab02bcf513fa1ed9d76e811a3cc
null
[ "LICENSE" ]
172
2.4
encommon
0.22.13
Enasis Network Common Library
# Enasis Network Common Library > This project has not released its first major version. Common classes and functions used in various public and private projects. <a href="https://pypi.org/project/encommon"><img src="https://enasisnetwork.github.io/encommon/badges/pypi.png"></a><br> <a href="https://enasisnetwork.github.io/encommon/validate/flake8.txt"><img src="https://enasisnetwork.github.io/encommon/badges/flake8.png"></a><br> <a href="https://enasisnetwork.github.io/encommon/validate/pylint.txt"><img src="https://enasisnetwork.github.io/encommon/badges/pylint.png"></a><br> <a href="https://enasisnetwork.github.io/encommon/validate/ruff.txt"><img src="https://enasisnetwork.github.io/encommon/badges/ruff.png"></a><br> <a href="https://enasisnetwork.github.io/encommon/validate/mypy.txt"><img src="https://enasisnetwork.github.io/encommon/badges/mypy.png"></a><br> <a href="https://enasisnetwork.github.io/encommon/validate/yamllint.txt"><img src="https://enasisnetwork.github.io/encommon/badges/yamllint.png"></a><br> <a href="https://enasisnetwork.github.io/encommon/validate/pytest.txt"><img src="https://enasisnetwork.github.io/encommon/badges/pytest.png"></a><br> <a href="https://enasisnetwork.github.io/encommon/validate/coverage.txt"><img src="https://enasisnetwork.github.io/encommon/badges/coverage.png"></a><br> <a href="https://enasisnetwork.github.io/encommon/validate/sphinx.txt"><img src="https://enasisnetwork.github.io/encommon/badges/sphinx.png"></a><br> ## Documentation Read [project documentation](https://enasisnetwork.github.io/encommon/sphinx) built using the [Sphinx](https://www.sphinx-doc.org/) project. Should you venture into the sections below you will be able to use the `sphinx` recipe to build documention in the `sphinx/html` directory. ## Projects using library - [Enasis Network Remote Connect](https://github.com/enasisnetwork/enconnect) - [Enasis Network Homie Automate](https://github.com/enasisnetwork/enhomie) - [Enasis Network Chatting Robie](https://github.com/enasisnetwork/enrobie) - [Enasis Network Orchestrations](https://github.com/enasisnetwork/orchestro) ## Installing the package Installing stable from the PyPi repository ``` pip install encommon ``` Installing latest from GitHub repository ``` pip install git+https://github.com/enasisnetwork/encommon ``` ## Quick start for local development Start by cloning the repository to your local machine. ``` git clone https://github.com/enasisnetwork/encommon.git ``` Set up the Python virtual environments expected by the Makefile. ``` make -s venv-create ``` ### Execute the linters and tests The comprehensive approach is to use the `check` recipe. This will stop on any failure that is encountered. ``` make -s check ``` However you can run the linters in a non-blocking mode. ``` make -s linters-pass ``` And finally run the various tests to validate the code and produce coverage information found in the `htmlcov` folder in the root of the project. ``` make -s pytest ``` ## Version management > :warning: Ensure that no changes are pending. 1. Rebuild the environment. ``` make -s check-revenv ``` 1. Update the [version.txt](encommon/version.txt) file. 1. Push to the `main` branch. 1. Create [repository](https://github.com/enasisnetwork/encommon) release. 1. Build the Python package.<br>Be sure no uncommited files in tree. ``` make -s pypackage ``` 1. Upload Python package to PyPi test. ``` make -s pypi-upload-test ``` 1. Upload Python package to PyPi prod. ``` make -s pypi-upload-prod ```
text/markdown
null
null
null
null
MIT
null
[ "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12" ]
[]
null
null
>=3.11
[]
[]
[]
[ "croniter", "cryptography", "jinja2", "netaddr", "pydantic", "python-dateutil", "pyyaml", "snaptime", "sqlalchemy" ]
[]
[]
[]
[ "Source, https://github.com/enasisnetwork/encommon", "Documentation, https://enasisnetwork.github.io/encommon/sphinx" ]
twine/6.2.0 CPython/3.12.6
2026-02-21T00:21:36.230188
encommon-0.22.13.tar.gz
76,060
6f/a8/f7273d9bdc92b65c95ec4caad4a619b0c1036b45f80f56018ab62cd037f3/encommon-0.22.13.tar.gz
source
sdist
null
false
68f17bd82efe34ec28649c8f8c79e633
deb2a5e6d4a55b09dffc6b34b688e55993008aa6f7acc5c15354114431f821a8
6fa8f7273d9bdc92b65c95ec4caad4a619b0c1036b45f80f56018ab62cd037f3
null
[ "LICENSE" ]
330
2.4
Bootmedian
1.1.2
A software to estimate medians through Bootstrapping
# Bootmedian Bootmedian estimates robust statistics (median, mean, sum, std) via bootstrapping and returns confidence intervals for the estimates. This package provides a small set of utilities centered on the `bootmedian()` function in `bootmedian/main.py` to compute bootstrapped statistics and an API to run bootstrap-based linear fits. **Highlights** - **Robust medians**: estimate medians using bootstrap resampling. - **Confidence intervals**: returns 1σ, 2σ and 3σ up/down limits for estimates. - **Multiple modes**: compute median, mean, std or sum via the `mode` argument. - **Weighted resampling**: supports sample weights in resampling. **Table of Contents** - **Overview**: short summary and behavior - **Installation**: dependencies and install - **Quickstart**: minimal usage examples - **API Reference**: main functions and parameters - **Notes & Contact**n+ **Overview** `bootmedian()` takes a 1-D array-like input and performs `nsimul` bootstrap resamples. By default it computes the median of each resample and returns the median of that distribution together with percentile-based confidence intervals for 1σ, 2σ and 3σ. NaN values in the input are ignored. Typical return (dictionary): - `median`: median of the bootstrap distribution (or mean/std/sum depending on `mode`). - `s1_up`, `s1_down`: 1σ upward and downward limits (percentiles). - `s2_up`, `s2_down`: 2σ limits. - `s3_up`, `s3_down`: 3σ limits. - `std1_up`, `std1_down`: optional percentiles of the raw sample (if `std` provided). - `sims`: the full bootstrap simulation array (useful for diagnostics). Installation Requirements: Python 3.8+ and the dependencies listed in `setup.py` / `pyproject.toml`. Install editable (dev) mode from the project root: ```bash pip install -e . ``` Or install via pip (from PyPI when published): ```bash pip install Bootmedian ``` Quickstart Example: compute the bootstrapped median and 1/2/3σ intervals for a sample: ```python import numpy as np from bootmedian import bootmedian data = np.array([1.0, 2.1, 2.3, np.nan, 3.5, 2.0]) result = bootmedian(data, nsimul=2000, errors=1, verbose=True) print(result) # -> dict with keys: 'median','s1_up','s1_down',...,'sims' ``` Using weights: ```python weights = np.array([1, 1, 2, 1, 1, 1.5]) result_w = bootmedian(data, nsimul=2000, weights=weights) ``` Change the statistic with `mode` ("median", "mean", "std", "sum"): ```python # Bootmedian Bootmedian estimates robust statistics (median, mean, sum, std) via bootstrapping and returns confidence intervals for the estimates. This package provides utilities centered on the `bootmedian()` function in `bootmedian/bootmedian.py` to compute bootstrapped statistics and an API to run bootstrap-based linear fits. Highlights - Robust medians: estimate medians using bootstrap resampling. - Confidence intervals: returns 1σ, 2σ and 3σ up/down limits for estimates. - Multiple modes: compute `median`, `mean`, `std` or `sum` via the `mode` argument. - Weighted resampling: supports sample weights in resampling. Table of contents - Overview - Installation - Quickstart - API reference - Notes & contact Overview `bootmedian()` takes a 1-D array-like input and performs `nsimul` bootstrap resamples. By default it computes the median of each resample and returns the median of that distribution together with percentile-based confidence intervals for 1σ, 2σ and 3σ. NaN values in the input are ignored. Typical return (dictionary) - `median`: median of the bootstrap distribution (or mean/std/sum depending on `mode`). - `s1_up`, `s1_down`: 1σ upward and downward limits (percentiles). - `s2_up`, `s2_down`: 2σ limits. - `s3_up`, `s3_down`: 3σ limits. - `std1_up`, `std1_down`: optional percentiles of the raw sample (if `std` provided). - `sims`: the full bootstrap simulation array (useful for diagnostics). Installation Requirements: Python 3.8+ and the dependencies listed in `setup.py` / `pyproject.toml`. Install in editable (development) mode from the project root: ```bash pip install -e . ``` Quickstart Example: compute the bootstrapped median and 1/2/3σ intervals for a sample: ```python import numpy as np from bootmedian import bootmedian data = np.array([1.0, 2.1, 2.3, np.nan, 3.5, 2.0]) result = bootmedian(data, nsimul=2000, errors=1, verbose=True) print(result) # -> dict with keys: 'median','s1_up','s1_down',...,'sims' ``` Using weights: ```python weights = np.array([1, 1, 2, 1, 1, 1.5]) result_w = bootmedian(data, nsimul=2000, weights=weights) ``` Change the statistic with `mode` ("median", "mean", "std", "sum"): ```python mean_result = bootmedian(data, nsimul=1500, mode="mean") ``` Bootstrap linear fit Use `bootfit(x, y, nsimul)` to obtain bootstrap distributions for slope (`m`) and intercept (`b`). The function returns a dictionary with medians and confidence percentiles for `m` and `b`. ```python from bootmedian import bootfit x = np.linspace(0, 10, 20) y = 2.3*x + 1.5 + np.random.normal(scale=0.5, size=x.size) fit = bootfit(x, y, nsimul=1000) print(fit['m_median'], fit['b_median']) ``` API reference (short) - `bootstrap_resample(X, weights=False, seed=None)` — Resamples an array-like `X` with optional `weights` (weighted sampling). Returns a flattened numpy array with one bootstrap resample. - `median_bootstrap(argument)` / `mean_bootstrap(argument)` / `sum_bootstrap(argument)` / `std_bootstrap(argument)` — Internal helpers used by `bootmedian` when running parallel workers. `argument` is a tuple/list: `(sample, weights, std)` where `std` is optional. - `boot_polyfit(x, y, seed)` — Performs a single resampled linear fit and returns `[slope, intercept]`. - `bootfit(x, y, nsimul, errors=1)` — Runs `nsimul` bootstrap fits (currently single-threaded loop with progress). Returns a dict with medians and percentile confidence intervals for `m` and `b`. - `bootmedian(sample_input, nsimul=1000, weights=False, errors=1, std=False, verbose=False, nthreads=7, mode="median")` — Main function; see docstring in code for parameter details. Notes & recommendations - The implementation uses `bottleneck` and `pandas.DataFrame.sample` for resampling. - For reproducibility you can set `numpy.random.seed(...)` before calling routines that internally use randomness. Some helper functions accept a `seed`. - `nsimul` controls accuracy vs runtime: start with a few hundred simulations, then increase to a few thousand if you need tighter percentiles. - If your input contains many NaNs ensure weights (if provided) align with non-NaN entries. Development & tests - See `setup.py` and `pyproject.toml` for declared dependencies. - A small example is available at `examples/simple_example.py` — run it after installing dependencies with `pip install -e .`. License & contact This project is released under the BSD license (see `setup.py` for metadata). Author: Alejandro S. Borlaff <a.s.borlaff@nasa.gov> --- If you'd like, I can add a short `examples/` notebook or a `requirements.txt` and run a quick sanity test to confirm imports.
text/markdown
Alejandro S. Borlaff
"Alejandro S. Borlaff" <a.s.borlaff@nasa.gov>
null
null
BSD
Statistics, Bootstrapping
[ "Development Status :: 3 - Alpha", "Topic :: Utilities", "License :: OSI Approved :: BSD License" ]
[]
https://github.com/Borlaff/bootmedian
null
null
[]
[]
[]
[ "numpy==1.26.4", "multiprocess", "bottleneck", "pandas", "tqdm", "astropy", "matplotlib", "miniutils" ]
[]
[]
[]
[ "Homepage, https://github.com/Borlaff/bootmedian" ]
twine/6.2.0 CPython/3.12.12
2026-02-21T00:20:32.290987
bootmedian-1.1.2.tar.gz
7,864
70/70/3ad916fcdd6e0d9032a3c1534de6d81993602670c7692b3b33aff5859cc6/bootmedian-1.1.2.tar.gz
source
sdist
null
false
716da1500032344d16e64f69569bd912
5f9fe29871bc801919d979c9e4971827953d2fb3918614d68eb00c7494b16d57
70703ad916fcdd6e0d9032a3c1534de6d81993602670c7692b3b33aff5859cc6
null
[ "LICENSE.md" ]
0
2.4
ai-edge-quantizer
0.5.0
A quantizer for advanced developers to quantize converted AI Edge models.
# AI Edge Quantizer A quantizer for advanced developers to quantize converted LiteRT models. It aims to facilitate advanced users to strive for optimal performance on resource demanding models (e.g., GenAI models). ## Build Status Build Type | Status | ----------- | --------------| Unit Tests (Linux) | [![](https://github.com/google-ai-edge/ai-edge-quantizer/actions/workflows/nightly_unittests.yml/badge.svg?branch=main)](https://github.com/google-ai-edge/ai-edge-quantizer/actions/workflows/nightly_unittests.yml) | Nightly Release | [![](https://github.com/google-ai-edge/ai-edge-quantizer/actions/workflows/nightly_release.yml/badge.svg?branch=main)](https://github.com/google-ai-edge/ai-edge-quantizer/actions/workflows/nightly_release.yml) | Nightly Colab | [![](https://github.com/google-ai-edge/ai-edge-quantizer/actions/workflows/nightly_colabs.yml/badge.svg?branch=main)](https://github.com/google-ai-edge/ai-edge-quantizer/actions/workflows/nightly_colabs.yml) | ## Installation ### Requirements and Dependencies * Python versions: 3.10, 3.11, 3.12, 3.13 * Operating system: Linux, MacOS * TensorFlow: [![tf-nightly](https://img.shields.io/badge/tf--nightly-latest-blue)](https://pypi.org/project/tf-nightly/) ### Install Nightly PyPi package: ```bash pip install ai-edge-quantizer-nightly ``` ## API Usage The quantizer requires two inputs: 1. An unquantized source LiteRT model (with FP32 data type in the FlatBuffer format with `.tflite` extension) 2. A quantization recipe (details below) and outputs a quantized LiteRT model that's ready for deployment on edge devices. ### Basic Usage In a nutshell, the quantizer works according to the following steps: 1. Instantiate a `Quantizer` class. This is the entry point to the quantizer's functionalities that the user accesses. 2. Load a desired quantization recipe (details in subsection). 3. Quantize (and save) the model. This is where most of the quantizer's internal logic works. ```python qt = quantizer.Quantizer("path/to/input/tflite") qt.load_quantization_recipe(recipe.dynamic_wi8_afp32()) qt.quantize().export_model("/path/to/output/tflite") ``` Please see the [getting started colab](colabs/getting_started.ipynb) for the simplest quick start guide on those 3 steps, and the [selective quantization colab](colabs/selective_quantization_isnet.ipynb) with more details on advanced features. #### LiteRT Model Please refer to the [LiteRT documentation](https://ai.google.dev/edge/litert) for ways to generate LiteRT models from Jax, PyTorch and TensorFlow. The input source model should be an FP32 (unquantized) model in the FlatBuffer format with `.tflite` extension. #### Quantization Recipe The user needs to specify a quantization recipe using AI Edge Quantizer's API to apply to the source model. The quantization recipe encodes all information on how a model is to be quantized, such as number of bits, data type, symmetry, scope name, etc. Essentially, a quantization recipe is defined as a collection of commands of the following type: _“Apply **Quantization Algorithm X** on **Operator Y** under **Scope Z** with **ConfigN**”._ For example: _\"**Uniformly quantize** the **FullyConnected op** under scope **'dense1/'** with **INT8 symmetric with Dynamic Quantization**"._ All the unspecified ops will be kept as FP32 (unquantized). The scope of an operator in TFLite is defined as the output tensor name of the op, which preserves the hierarchical model information from the source model (e.g., scope in TF). The best way to obtain scope name is by visualizing the model with [Model Explorer](https://ai.google.dev/edge/model-explorer). Currently, there are three ways to quantize an operator: * **dynamic quantization (recommended)**: weights are quantized while activations remain in a float format and are not processed by AI Edge Quantizer (AEQ). The runtime kernel handles the on-the-fly quantization of these activations, as identified by `compute_precision=integer` and `explicit_dequantize=False`. * Pros: reduced model size and memory usage. Latency improvement due to integer computation. No sample data requirement (calibration). * Cons: on-the-fly quantization of activation tensors can affect model quality. Not supported in all hardware (e.g., some GPU and NPU). * **weight only quantization**: only model weights are quantized, not activations. The actual operation (op) computation remains in float. The quantized weight is explicitly dequantized before being fed into the op, by inserting a dequantize op between the quantized weight and the consuming op. To enable this, `compute_precision` will be set to `float` and `explicit_dequantize` to `True`. * Pros: reduced model size and memory usage. No sample data requirement (calibration). Usually has the best model quality. * Cons: no latency benefit (may be worse) due to float computation with explicit dequantization. * **static quantization**: both weights and activations are quantized. This requires a calibration phase to estimate quantization parameters of runtime tensors (activations). * Pros: reduced model size, memory usage, and latency. * Cons: requires sample data for calibration. Imposing static quantization parameters (derived from calibration) on runtime tensors can compromise quality. Generally, we recommend dynamic quantization for CPU/GPU deployment and static quantization for NPU deployment. We include commonly used recipes in [recipe.py](ai_edge_quantizer/recipe.py). This is demonstrated in the [getting started colab](colabs/getting_started.ipynb) example. Advanced users can build their own recipe through the quantizer API. #### Deployment Please refer to the [LiteRT deployment documentation](https://ai.google.dev/edge/litert/inference) for ways to deploy a quantized LiteRT model. ### Advanced Recipes There are many ways the user can configure and customize the quantization recipe beyond using a template in [recipe.py](ai_edge_quantizer/recipe.py). For example, the user can configure the recipe to achieve these features: * Selective quantization (exclude selected ops from being quantized) * Flexible mixed scheme quantization (mixture of different precision, compute precision, scope, op, config, etc) * 4-bit weight quantization The [selective quantization colab](colabs/selective_quantization_isnet.ipynb) shows some of these more advanced features. For specifics of the recipe schema, please refer to the `OpQuantizationRecipe` in [recipe_manager.py]. For advanced usage involving mixed quantization, the following API may be useful: * Use `Quantizer:load_quantization_recipe()` in [quantizer.py](ai_edge_quantizer/quantizer.py) to load a custom recipe. * Use `Quantizer:update_quantization_recipe()` in [quantizer.py](ai_edge_quantizer/quantizer.py) to extend or override specific parts of the recipe. ### Operator coverage The table below outlines the allowed configurations for available recipes. | | | | | | | | | | | | --- | --- | --- | --- | --- | --- |--- |--- |--- |--- | | **Config** | | DYNAMIC_WI8_AFP32 | DYNAMIC_WI4_AFP32 | STATIC_WI8_AI16 | STATIC_WI4_AI16 | STATIC_WI8_AI8 | STATIC_WI4_AI8 | WEIGHTONLY_WI8_AFP32 | WEIGHTONLY_WI4_AFP32 | |activation| num\_bits | None | None | 16 | 16 | 8 | 8 | None | None | | | symmetric |None | None | TRUE | TRUE | [TRUE, FALSE] | [TRUE, FALSE] | None | None | | | granularity |None | None | TENSORWISE | TENSORWISE | TENSORWISE | TENSORWISE | None | None | | | dtype| None | None |INT | INT | INT | INT | None | None | | weight | num\_bits | 8 | 4 | 8 | 4 | 8 | 4 | 8 | 4 | | | symmetric | TRUE | TRUE | TRUE | TRUE | TRUE | TRUE | [TRUE, FALSE] | [TRUE, FALSE] | | | granularity | \[CHANNELWISE, TENSORWISE\] | \[CHANNELWISE, TENSORWISE\] | \[CHANNELWISE, TENSORWISE\] | \[CHANNELWISE, TENSORWISE\] | \[CHANNELWISE, TENSORWISE\] | \[CHANNELWISE, TENSORWISE\] | \[CHANNELWISE, TENSORWISE\] | \[CHANNELWISE, TENSORWISE\] | | | dtype | INT | INT | INT | INT | INT | INT | INT | INT | | explicit\_dequantize | | FALSE | FALSE | FALSE | FALSE | FALSE | FALSE | TRUE | TRUE | | compute\_precision || INTEGER | INTEGER | INTEGER | INTEGER | INTEGER | INTEGER | FLOAT | FLOAT | **Operators Supporting Quantization** | | | | | | | | | | | --- | --- | --- | --- | --- | --- |--- |--- |--- | | **Config** | DYNAMIC_WI8_AFP32 | DYNAMIC_WI4_AFP32 | STATIC_WI8_AI16 | STATIC_WI4_AI16 | STATIC_WI8_AI8 | STATIC_WI4_AI8 | WEIGHTONLY_WI8_AFP32 | WEIGHTONLY_WI4_AFP32 | |FULLY_CONNECTED |<div align="center"> &check; </div>|<div align="center"> &check; </div>|<div align="center"> &check; </div>|<div align="center"> &check; </div>|<div align="center"> &check; </div>|<div align="center"> &check; </div>|<div align="center"> &check; </div>|<div align="center"> &check; </div>| |CONV_2D |<div align="center"> &check; </div>| |<div align="center"> &check; </div>|<div align="center"> &check; </div>|<div align="center"> &check; </div>|<div align="center"> &check; </div>|<div align="center"> &check; </div>| | |BATCH_MATMUL |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | |EMBEDDING_LOOKUP |<div align="center"> &check; </div>|<div align="center"> &check; </div>|<div align="center"> &check; </div>| |<div align="center"> &check; </div>|<div align="center"> &check; </div>|<div align="center"> &check; </div>| | |DEPTHWISE_CONV_2D|<div align="center"> &check; </div>| |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | |AVERAGE_POOL_2D | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |RESHAPE | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |SOFTMAX | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |TANH | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |TRANSPOSE | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |GELU | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |ADD | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |CONV_2D_TRANSPOSE|<div align="center"> &check; </div>| |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |SUB | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |MUL | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |MEAN | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |RSQRT | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |CONCATENATION | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |STRIDED_SLICE | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |SPLIT | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |LOGISTIC | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |SLICE | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |SELECT | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |SELECT_V2 | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |SUM | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |PAD | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |PADV2 | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |MIRROR_PAD | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |SQUARED_DIFFERENCE | | | | |<div align="center"> &check; </div>| | | | |MAX_POOL_2D | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |RESIZE_BILINEAR | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |RESIZE_NEAREST_NEIGHBOR| | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |GATHER_ND | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |PACK | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |UNPACK | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |DIV | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |SQRT | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |GATHER | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |HARD_SWISH | | | | |<div align="center"> &check; </div>| | | | |MAXIMUM | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |REDUCE_MIN | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |EQUAL | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |NOT_EQUAL | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | | |SPACE_TO_DEPTH | | | | |<div align="center"> &check; </div>| | | | |RELU | | |<div align="center"> &check; </div>| |<div align="center"> &check; </div>| | | |
text/markdown
null
null
null
null
Apache-2.0
On-Device ML, AI, Google, TFLite, Quantization, LLMs, GenAI
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Scientific/Engineering", "Topic :: Scientific/Engineering :: Mathematics", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules" ]
[]
https://github.com/google-ai-edge/ai-edge-quantizer
null
>=3.10
[]
[]
[]
[ "absl-py", "immutabledict", "numpy", "ml_dtypes", "ai-edge-litert-nightly" ]
[]
[]
[]
[ "Homepage, https://github.com/google-ai-edge/ai-edge-quantizer" ]
twine/6.2.0 CPython/3.11.14
2026-02-21T00:18:36.025863
ai_edge_quantizer-0.5.0-py3-none-any.whl
388,367
49/9e/bc734b5b51bd7dfe2b1d4ec23805e23d8212a7fa6082d82adc674ded565d/ai_edge_quantizer-0.5.0-py3-none-any.whl
py3
bdist_wheel
null
false
00ef9246fd60ad767ddcd9efe9acb001
ac563690bae65fac17ee40980950e6e8c105b47cd5884816f141326c754c96bf
499ebc734b5b51bd7dfe2b1d4ec23805e23d8212a7fa6082d82adc674ded565d
null
[ "LICENSE" ]
89
2.4
sdv
1.34.1
Generate synthetic data for single table, multi table and sequential data
<div align="center"> <br/> <p align="center"> <i>This repository is part of <a href="https://sdv.dev">The Synthetic Data Vault Project</a>, a project from <a href="https://datacebo.com">DataCebo</a>.</i> </p> [![Dev Status](https://img.shields.io/badge/Dev%20Status-5%20--%20Production%2fStable-green)](https://pypi.org/search/?c=Development+Status+%3A%3A+5+-+Production%2FStable) [![PyPi Shield](https://img.shields.io/pypi/v/SDV.svg)](https://pypi.python.org/pypi/SDV) [![Unit Tests](https://github.com/sdv-dev/SDV/actions/workflows/unit.yml/badge.svg?branch=main)](https://github.com/sdv-dev/SDV/actions/workflows/unit.yml?query=branch%3Amain) [![Integration Tests](https://github.com/sdv-dev/SDV/actions/workflows/integration.yml/badge.svg?branch=main)](https://github.com/sdv-dev/SDV/actions/workflows/integration.yml?query=branch%3Amain) [![Coverage Status](https://codecov.io/gh/sdv-dev/SDV/branch/main/graph/badge.svg)](https://codecov.io/gh/sdv-dev/SDV) [![Downloads](https://static.pepy.tech/personalized-badge/sdv?period=total&units=international_system&left_color=grey&right_color=blue&left_text=Downloads)](https://pepy.tech/project/sdv) [![Colab](https://img.shields.io/badge/Tutorials-Try%20now!-orange?logo=googlecolab)](https://docs.sdv.dev/sdv/demos) [![Forum](https://img.shields.io/badge/Forum-Join%20now!-36C5F0)](https://forum.datacebo.com) <div align="left"> <br/> <p align="center"> <a href="https://github.com/sdv-dev/SDV"> <img align="center" width=40% src="https://github.com/sdv-dev/SDV/blob/stable/docs/images/SDV-logo.png"></img> </a> </p> </div> </div> # Overview The **Synthetic Data Vault** (SDV) is a Python library designed to be your one-stop shop for creating tabular synthetic data. The SDV uses a variety of machine learning algorithms to learn patterns from your real data and emulate them in synthetic data. ## Features :brain: **Create synthetic data using machine learning.** The SDV offers multiple models, ranging from classical statistical methods (GaussianCopula) to deep learning methods (CTGAN). Generate data for single tables, multiple connected tables or sequential tables. :bar_chart: **Evaluate and visualize data.** Compare the synthetic data to the real data against a variety of measures. Diagnose problems and generate a quality report to get more insights. :arrows_counterclockwise: **Preprocess, anonymize and define constraints.** Control data processing to improve the quality of synthetic data, choose from different types of anonymization and define business rules in the form of logical constraints. | Important Links | | | --------------------------------------------- | ----------------------------------------------------------------------------------------------------| | [![][Colab Logo] **Tutorials**][Tutorials] | Get some hands-on experience with the SDV. Launch the tutorial notebooks and run the code yourself. | | :book: **[Docs]** | Learn how to use the SDV library with user guides and API references. | | :orange_book: **[Blog]** | Get more insights about using the SDV, deploying models and our synthetic data community. | | :busts_in_silhouette: **[DataCebo Forum]** | Discuss SDV features, ask questions, and receive help . | | :computer: **[Website]** | Check out the SDV website for more information about the project. | [Website]: https://sdv.dev [Blog]: https://datacebo.com/blog [Docs]: https://bit.ly/sdv-docs [Repository]: https://github.com/sdv-dev/SDV [License]: https://github.com/sdv-dev/SDV/blob/main/LICENSE [Development Status]: https://pypi.org/search/?c=Development+Status+%3A%3A+5+-+Production%2FStable [DataCebo Forum]: https://forum.datacebo.com [Colab Logo]: https://github.com/sdv-dev/SDV/blob/stable/docs/images/google_colab.png [Tutorials]: https://docs.sdv.dev/sdv/demos # Install The SDV is publicly available under the [Business Source License](https://github.com/sdv-dev/SDV/blob/main/LICENSE). Install SDV using pip or conda. We recommend using a virtual environment to avoid conflicts with other software on your device. ```bash pip install sdv ``` ```bash conda install -c pytorch -c conda-forge sdv ``` # Getting Started Load a demo dataset to get started. This dataset is a single table describing guests staying at a fictional hotel. ```python from sdv.datasets.demo import download_demo real_data, metadata = download_demo( modality='single_table', dataset_name='fake_hotel_guests') ``` ![Single Table Metadata Example](https://github.com/sdv-dev/SDV/blob/stable/docs/images/Single-Table-Metadata-Example.png) The demo also includes **metadata**, a description of the dataset, including the data types in each column and the primary key (`guest_email`). ## Synthesizing Data Next, we can create an **SDV synthesizer**, an object that you can use to create synthetic data. It learns patterns from the real data and replicates them to generate synthetic data. Let's use the [GaussianCopulaSynthesizer](https://docs.sdv.dev/sdv/single-table-data/modeling/synthesizers/gaussiancopulasynthesizer). ```python from sdv.single_table import GaussianCopulaSynthesizer synthesizer = GaussianCopulaSynthesizer(metadata) synthesizer.fit(data=real_data) ``` And now the synthesizer is ready to create synthetic data! ```python synthetic_data = synthesizer.sample(num_rows=500) ``` The synthetic data will have the following properties: - **Sensitive columns are fully anonymized.** The email, billing address and credit card number columns contain new data so you don't expose the real values. - **Other columns follow statistical patterns.** For example, the proportion of room types, the distribution of check in dates and the correlations between room rate and room type are preserved. - **Keys and other relationships are intact.** The primary key (guest email) is unique for each row. If you have multiple tables, the connection between a primary and foreign keys makes sense. ## Evaluating Synthetic Data The SDV library allows you to evaluate the synthetic data by comparing it to the real data. Get started by generating a quality report. ```python from sdv.evaluation.single_table import evaluate_quality quality_report = evaluate_quality( real_data, synthetic_data, metadata) ``` ``` Generating report ... (1/2) Evaluating Column Shapes: |████████████████| 9/9 [00:00<00:00, 1133.09it/s]| Column Shapes Score: 89.11% (2/2) Evaluating Column Pair Trends: |██████████████████████████████████████████| 36/36 [00:00<00:00, 502.88it/s]| Column Pair Trends Score: 88.3% Overall Score (Average): 88.7% ``` This object computes an overall quality score on a scale of 0 to 100% (100 being the best) as well as detailed breakdowns. For more insights, you can also visualize the synthetic vs. real data. ```python from sdv.evaluation.single_table import get_column_plot fig = get_column_plot( real_data=real_data, synthetic_data=synthetic_data, column_name='amenities_fee', metadata=metadata ) fig.show() ``` ![Real vs. Synthetic Data](https://github.com/sdv-dev/SDV/blob/stable/docs/images/Real-vs-Synthetic-Evaluation.png) # What's Next? Using the SDV library, you can synthesize single table, multi table and sequential data. You can also customize the full synthetic data workflow, including preprocessing, anonymization and adding constraints. To learn more, visit the [SDV Demo page](https://docs.sdv.dev/sdv/demos). # Credits Thank you to our team of contributors who have built and maintained the SDV ecosystem over the years! [View Contributors](https://github.com/sdv-dev/SDV/graphs/contributors) ## Citation If you use SDV for your research, please cite the following paper: *Neha Patki, Roy Wedge, Kalyan Veeramachaneni*. [The Synthetic Data Vault](https://dai.lids.mit.edu/wp-content/uploads/2018/03/SDV.pdf). [IEEE DSAA 2016](https://ieeexplore.ieee.org/document/7796926). ``` @inproceedings{ SDV, title={The Synthetic data vault}, author={Patki, Neha and Wedge, Roy and Veeramachaneni, Kalyan}, booktitle={IEEE International Conference on Data Science and Advanced Analytics (DSAA)}, year={2016}, pages={399-410}, doi={10.1109/DSAA.2016.49}, month={Oct} } ``` --- <div align="center"> <a href="https://datacebo.com"><picture> <source media="(prefers-color-scheme: dark)" srcset="https://github.com/sdv-dev/SDV/blob/stable/docs/images/datacebo-logo-dark-mode.png"> <img align="center" width=40% src="https://github.com/sdv-dev/SDV/blob/stable/docs/images/datacebo-logo.png"></img> </picture></a> </div> <br/> <br/> [The Synthetic Data Vault Project](https://sdv.dev) was first created at MIT's [Data to AI Lab]( https://dai.lids.mit.edu/) in 2016. After 4 years of research and traction with enterprise, we created [DataCebo](https://datacebo.com) in 2020 with the goal of growing the project. Today, DataCebo is the proud developer of SDV, the largest ecosystem for synthetic data generation & evaluation. It is home to multiple libraries that support synthetic data, including: * 🔄 Data discovery & transformation. Reverse the transforms to reproduce realistic data. * 🧠 Multiple machine learning models -- ranging from Copulas to Deep Learning -- to create tabular, multi table and time series data. * 📊 Measuring quality and privacy of synthetic data, and comparing different synthetic data generation models. [Get started using the SDV package](https://bit.ly/sdv-docs) -- a fully integrated solution and your one-stop shop for synthetic data. Or, use the standalone libraries for specific needs.
text/markdown
null
"DataCebo, Inc." <info@sdv.dev>
null
null
null
sdv, synthetic-data, synthetic-data-generation, timeseries, single-table, multi-table
[ "Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Natural Language :: English", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14", "Topic :: Scientific/Engineering :: Artificial Intelligence" ]
[]
null
null
<3.15,>=3.9
[]
[]
[]
[ "boto3<2.0.0,>=1.28", "botocore<2.0.0,>=1.31", "cloudpickle>=2.1.0; python_version < \"3.14\"", "cloudpickle>=3.1.1; python_version >= \"3.14\"", "graphviz>=0.13.2", "numpy>=1.22.2; python_version < \"3.10\"", "numpy>=1.24.0; python_version >= \"3.10\" and python_version < \"3.12\"", "numpy>=1.26.0; python_version >= \"3.12\" and python_version < \"3.13\"", "numpy>=2.1.0; python_version >= \"3.13\" and python_version < \"3.14\"", "numpy>=2.3.2; python_version >= \"3.14\"", "pandas<3,>=1.4.0; python_version < \"3.11\"", "pandas<3,>=1.5.0; python_version >= \"3.11\" and python_version < \"3.12\"", "pandas<3,>=2.1.1; python_version >= \"3.12\" and python_version < \"3.13\"", "pandas<3,>=2.2.3; python_version >= \"3.13\" and python_version < \"3.14\"", "pandas<3,>=2.3.3; python_version >= \"3.14\"", "tqdm>=4.29", "copulas>=0.12.1; python_version < \"3.14\"", "copulas>=0.14.0; python_version >= \"3.14\"", "ctgan>=0.11.1; python_version < \"3.14\"", "ctgan>=0.12.0; python_version >= \"3.14\"", "deepecho>=0.7.0; python_version < \"3.14\"", "deepecho>=0.8.0; python_version >= \"3.14\"", "rdt>=1.18.2; python_version < \"3.14\"", "rdt>=1.20.0; python_version >= \"3.14\"", "sdmetrics>=0.21.0; python_version < \"3.14\"", "sdmetrics>=0.26.0; python_version >= \"3.14\"", "platformdirs>=4.0", "pyyaml>=6.0.1", "pandas[excel]; extra == \"excel\"", "sdv[excel]; extra == \"test\"", "pytest>=3.4.2; extra == \"test\"", "pytest-cov>=2.6.0; extra == \"test\"", "pytest-rerunfailures<17,>=10.3; extra == \"test\"", "jupyter<2,>=1.0.0; extra == \"test\"", "pytest-runner>=2.11.1; extra == \"test\"", "tomli<3,>=2.0.0; extra == \"test\"", "google-api-python-client; extra == \"test\"", "google-auth; extra == \"test\"", "google-auth-oauthlib; extra == \"test\"", "requests; extra == \"test\"", "pyarrow; extra == \"test\"", "gitpython; extra == \"test\"", "slack-sdk<4.0,>=3.23; extra == \"test\"", "pomegranate<1,>=0.15; extra == \"pomegranate\"", "sdv[test]; extra == \"dev\"", "build<2,>=1.0.0; extra == \"dev\"", "bump-my-version>=0.18.3; extra == \"dev\"", "pip>=9.0.1; extra == \"dev\"", "watchdog<5,>=1.0.1; extra == \"dev\"", "docutils<1,>=0.12; extra == \"dev\"", "nbsphinx<1,>=0.5.0; extra == \"dev\"", "sphinx_toolbox<4,>=2.5; extra == \"dev\"", "Sphinx<8,>=3; extra == \"dev\"", "pydata-sphinx-theme<1; extra == \"dev\"", "markupsafe<3; extra == \"dev\"", "lxml_html_clean<0.5; extra == \"dev\"", "sphinx-reredirects; extra == \"dev\"", "Jinja2<4,>=2; extra == \"dev\"", "ruff<1,>=0.4.5; extra == \"dev\"", "twine>=1.10.0; extra == \"dev\"", "wheel>=0.30.0; extra == \"dev\"", "coverage<8,>=4.5.12; extra == \"dev\"", "invoke; extra == \"dev\"", "rundoc<0.5,>=0.4.3; extra == \"readme\"" ]
[]
[]
[]
[ "Source Code, https://github.com/sdv-dev/SDV/", "Issue Tracker, https://github.com/sdv-dev/SDV/issues", "Changes, https://github.com/sdv-dev/SDV/blob/main/HISTORY.md", "Twitter, https://twitter.com/sdv_dev", "Chat, https://forum.datacebo.com" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:17:37.033220
sdv-1.34.1.tar.gz
173,389
7c/54/ea499d3080ff89eec1c118c2543ab7f63de9b3cd0cff85ee92c0a12d7fa0/sdv-1.34.1.tar.gz
source
sdist
null
false
86a479705312abf45d93b1922e480bd5
77cc0c77f38fa7e56a0ec15b20f4c52ae69b3b7e2a9f649a5a57ace6d64c01b0
7c54ea499d3080ff89eec1c118c2543ab7f63de9b3cd0cff85ee92c0a12d7fa0
BUSL-1.1
[ "LICENSE" ]
659
2.3
rush-py
6.5.1
Python client for QDX's Rush platform
# rush-py: Rush Python Client ## Installation Install from PyPI: ```bash pip install rush-py ``` If you manage dependencies with `uv`, add it to your project (updates `pyproject.toml` and `uv.lock`): ```bash uv add rush-py ``` ### Using in Your Project Add to your `pyproject.toml`: ```toml [project] dependencies = [ "rush-py", ] ``` ## Rush Setup Use environment variables to configure access: - `RUSH_TOKEN`: Put your token's value here - `RUSH_PROJECT`: Put your project's UUID value here; can find it in the URL once selecting a project in the Rush UI - `RUSH_ENDPOINT`: Use this to choose between staging and prod; if omitted, defaults to prod You can also put `RUSH_TOKEN` and `RUSH_PROJECT` in a `.env` file instead of exporting them in every terminal session. rush-py looks for a `.env` file in the current working directory first, then falls back to `~/.rush/.env`. Environment variables always take priority over `.env` values. ``` # .env RUSH_TOKEN=your-token-here RUSH_PROJECT=your-project-id-here ``` ## Quick Start ```python from pathlib import Path from rush import exess from rush.client import collect_run topology_path = Path.cwd() / "thrombin_1c_t.json" # For energy, the only mandatory argument is the Topology result = exess.energy(topology_path, collect=True) exess.save_energy_outputs(result) ``` Outputs are saved under `<workspace_dir>/<PROJECT_ID>/` (default: current working directory). To customize the workspace location, call `rush.client.set_opts(workspace_dir=Path("..."))`. ```python # For interaction_energy, second argument is reference fragment result = exess.interaction_energy(topology_path, 1, collect=True) # For chelpg, charges are extracted from the HDF5 output and returned as a list output_info, charges = exess.chelpg(topology_path, collect=True) # QMMM requires Residues too md_topology_path = "./6a5j_t.json" md_residues_path = "./6a5j_r.json" # Without `collect=True`, the run takes place asynchronously, and a run ID is returned id = exess.qmmm( md_topology_path, md_residues_path, n_timesteps=500, qm_fragments=[0], free_atoms=[0], ) # The output for qmmm is a geometries json; can swap into a Topology's geometry field result = collect_run(id) # Get the full list of parameters and default arguments for a function help(exess.energy) help(exess.interaction_energy) help(exess.chelpg) help(exess.qmmm) ``` See the [docs](https://talo.github.io/rush-py) for more information! ## Development You can develop this project using `pip` + `venv`, or `uv`. ### With `pip` + `venv` ```bash git clone git@github.com:talo/rush-py.git cd rush-py python -m venv .venv source .venv/bin/activate pip install -e . ``` ### With `uv` ```bash git clone git@github.com:talo/rush-py.git cd rush-py uv sync source .venv/bin/activate ``` See the Terms of Service for use of the underlying Rush software at https://qdx.co/terms.
text/markdown
Sean Laguna
Sean Laguna <seanlaguna@qdx.co>
null
null
Apache-2.0
computational-chemistry, quantum-chemistry, HPC, molecular-modeling, protein-folding, EXESS, NN-xTB, PBSA
[ "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.12", "Intended Audience :: Healthcare Industry", "Intended Audience :: Science/Research", "Topic :: Scientific/Engineering :: Bio-Informatics", "Topic :: Scientific/Engineering :: Chemistry", "Topic :: System :: Distributed Computing" ]
[]
null
null
>=3.12
[]
[]
[]
[ "gql~=4.0", "h5py~=3.14", "rdkit~=2025.9.2", "matplotlib~=3.10", "networkx~=3.6", "requests~=2.32", "requests-toolbelt~=1.0", "zstandard~=0.23" ]
[]
[]
[]
[ "Homepage, https://rush.cloud", "Documentation, https://talo.github.io/rush-py", "Source, https://github.com/talo/rush-py", "Issues, https://github.com/talo/rush-py/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:17:21.840394
rush_py-6.5.1.tar.gz
60,852
26/45/466741b2113d1a166f7c4a74604d6555c3f4de2038801e4dd4c2eed2909e/rush_py-6.5.1.tar.gz
source
sdist
null
false
89b4c944e321250f7656f76d133a5a44
fdf8e32afa61a14680c844b8cd6bf9c6f2eed711518474ab9d22890e6f319320
2645466741b2113d1a166f7c4a74604d6555c3f4de2038801e4dd4c2eed2909e
null
[]
186
2.4
vws-python-mock
2026.2.21
A mock for the Vuforia Web Services (VWS) API.
|Build Status| |PyPI| VWS Mock ======== .. contents:: :local: Mock for the Vuforia Web Services (VWS) API and the Vuforia Web Query API. Mocking calls made to Vuforia with Python ``requests`` ------------------------------------------------------ Using the mock redirects requests to Vuforia made with `requests`_ to an in-memory implementation. .. code-block:: shell pip install vws-python-mock This requires Python |minimum-python-version|\+. .. code-block:: python """Make a request to the Vuforia Web Services API mock.""" import requests from mock_vws import MockVWS from mock_vws.database import CloudDatabase with MockVWS() as mock: database = CloudDatabase() mock.add_cloud_database(cloud_database=database) # This will use the Vuforia mock. requests.get(url="https://vws.vuforia.com/summary", timeout=30) By default, an exception will be raised if any requests to unmocked addresses are made. .. _requests: https://pypi.org/project/requests/ Using Docker to mock calls to Vuforia from any language ------------------------------------------------------- It is possible run a Mock VWS instance using Docker containers. This allows you to run tests against a mock VWS instance regardless of the language or tooling you are using. See the `the instructions <https://vws-python.github.io/vws-python-mock/docker.html>`__ for how to do this. Full documentation ------------------ See the `full documentation <https://vws-python.github.io/vws-python-mock/>`__. This includes details on how to use the mock, options, and details of the differences between the mock and the real Vuforia Web Services. .. |Build Status| image:: https://github.com/VWS-Python/vws-python-mock/actions/workflows/test.yml/badge.svg?branch=main :target: https://github.com/VWS-Python/vws-python-mock/actions .. |PyPI| image:: https://badge.fury.io/py/VWS-Python-Mock.svg :target: https://badge.fury.io/py/VWS-Python-Mock .. |minimum-python-version| replace:: 3.13
text/x-rst
null
Adam Dangoor <adamdangoor@gmail.com>
null
null
null
client, fake, mock, vuforia, vws
[ "Development Status :: 5 - Production/Stable", "Environment :: Web Environment", "Framework :: Pytest", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.13" ]
[]
null
null
>=3.13
[]
[]
[]
[ "beartype>=0.22.9", "flask>=3.0.3", "numpy>=1.26.4", "pillow>=11.0.0", "piq>=0.8.0", "pydantic-settings>=2.6.1", "requests>=2.32.3", "responses>=0.25.3", "torch>=2.5.1", "torchmetrics>=1.5.1", "torchvision>=0.20.1", "tzdata; sys_platform == \"win32\"", "vws-auth-tools>=2024.7.12", "werkzeug>=3.1.2", "actionlint-py==1.7.11.24; extra == \"dev\"", "check-manifest==0.51; extra == \"dev\"", "check-wheel-contents==0.6.3; extra == \"dev\"", "coverage==7.13.4; extra == \"dev\"", "deptry==0.24.0; extra == \"dev\"", "dirty-equals==0.11; extra == \"dev\"", "doc8==2.0.0; extra == \"dev\"", "doccmd==2026.2.15; extra == \"dev\"", "docker==7.1.0; extra == \"dev\"", "enum-tools[sphinx]==0.13.0; extra == \"dev\"", "freezegun==1.5.5; extra == \"dev\"", "furo==2025.12.19; extra == \"dev\"", "interrogate==1.7.0; extra == \"dev\"", "mypy[faster-cache]==1.19.1; extra == \"dev\"", "mypy-strict-kwargs==2026.1.12; extra == \"dev\"", "prek==0.3.3; extra == \"dev\"", "pydocstringformatter==0.7.5; extra == \"dev\"", "pydocstyle==6.3; extra == \"dev\"", "pylint[spelling]==4.0.4; extra == \"dev\"", "pylint-per-file-ignores==3.2.0; extra == \"dev\"", "pyproject-fmt==2.16.1; extra == \"dev\"", "pyrefly==0.53.0; extra == \"dev\"", "pyright==1.1.408; extra == \"dev\"", "pyroma==5.0.1; extra == \"dev\"", "pytest==9.0.2; extra == \"dev\"", "pytest-retry==1.7.0; extra == \"dev\"", "pytest-xdist==3.8.0; extra == \"dev\"", "python-dotenv==1.2.1; extra == \"dev\"", "pyyaml==6.0.3; extra == \"dev\"", "requests-mock-flask==2026.2.16; extra == \"dev\"", "ruff==0.15.1; extra == \"dev\"", "shellcheck-py==0.11.0.1; extra == \"dev\"", "shfmt-py==3.12.0.2; extra == \"dev\"", "sphinx==8.2.3; extra == \"dev\"", "sphinx-copybutton==0.5.2; extra == \"dev\"", "sphinx-lint==1.0.2; extra == \"dev\"", "sphinx-paramlinks==0.6; extra == \"dev\"", "sphinx-pyproject==0.3.0; extra == \"dev\"", "sphinx-substitution-extensions==2026.1.12; extra == \"dev\"", "sphinx-toolbox==4.1.2; extra == \"dev\"", "sphinxcontrib-httpdomain==2.0.0; extra == \"dev\"", "sphinxcontrib-spelling==8.0.2; extra == \"dev\"", "sybil==9.3.0; extra == \"dev\"", "tenacity==9.1.4; extra == \"dev\"", "ty==0.0.17; extra == \"dev\"", "types-docker==7.1.0.20260109; extra == \"dev\"", "types-pyyaml==6.0.12.20250915; extra == \"dev\"", "types-requests==2.32.4.20260107; extra == \"dev\"", "urllib3==2.6.3; extra == \"dev\"", "vulture==2.14; extra == \"dev\"", "vws-python==2026.2.15; extra == \"dev\"", "vws-test-fixtures==2023.3.5; extra == \"dev\"", "vws-web-tools==2026.2.20; extra == \"dev\"", "yamlfix==1.19.1; extra == \"dev\"", "zizmor==1.22.0; extra == \"dev\"", "check-wheel-contents==0.6.3; extra == \"release\"" ]
[]
[]
[]
[ "Documentation, https://vws-python.github.io/vws-python-mock/", "Source, https://github.com/VWS-Python/vws-python-mock" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:17:16.080477
vws_python_mock-2026.2.21.tar.gz
158,586
a8/c6/73e35a2c9390c2db650083c22db86e1b160b78751130d410d08a1667306d/vws_python_mock-2026.2.21.tar.gz
source
sdist
null
false
960acc1893c13e9aea71b66043624bd0
173d901643b8ea9fc643aad3289330d4bb3a3525a11eeb3c89b268e7c4d6b26e
a8c673e35a2c9390c2db650083c22db86e1b160b78751130d410d08a1667306d
MIT
[ "LICENSE" ]
523
2.4
phlo-dlt
0.1.1
DLT ingestion engine capability plugin for Phlo
DLT ingestion engine capability plugin for Phlo.
text/plain
null
Phlo Team <team@phlo.dev>
null
null
MIT
null
[]
[]
null
null
>=3.11
[]
[]
[]
[ "phlo>=0.1.0", "dlt>=1.18.2", "pandas>=2.3.3", "pandera>=0.26.1", "pyarrow>=21.0.0", "pyiceberg>=0.10.0", "python-ulid>=2.0.0", "pytest>=7.0; extra == \"dev\"", "ruff>=0.1.0; extra == \"dev\"" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:17:01.403989
phlo_dlt-0.1.1.tar.gz
25,917
79/31/56588b328cf775457811e7480ee1dbfd81017be5bde45e0e77fac1c45000/phlo_dlt-0.1.1.tar.gz
source
sdist
null
false
c9f4c01e729ac7e6d1aff2c0eb6430ef
4c54f3a6f473973a960eb1bb39ef6af531a4338f501c89a8067b54585bbca842
793156588b328cf775457811e7480ee1dbfd81017be5bde45e0e77fac1c45000
null
[]
190
2.4
phlo-dbt
0.1.1
dbt integration capability plugin for Phlo
dbt integration capability plugin for Phlo.
text/plain
null
Phlo Team <team@phlo.dev>
null
null
MIT
null
[]
[]
null
null
>=3.11
[]
[]
[]
[ "phlo>=0.1.0", "dbt-core>=1.10.8", "dbt-trino>=1.9.3", "dbt-postgres>=1.9.1", "pytest>=7.0; extra == \"dev\"", "ruff>=0.1.0; extra == \"dev\"", "phlo-dagster>=0.1.0; extra == \"dagster\"", "phlo-quality>=0.1.0; extra == \"quality\"" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:17:00.460204
phlo_dbt-0.1.1.tar.gz
30,270
91/60/8263a06a0b481fe41fed07828b0e4846fbdde761b612d8e403bb4317f5ec/phlo_dbt-0.1.1.tar.gz
source
sdist
null
false
340721e92bfcd96adc78b7c55047b31f
a8c70f5c48513d014660290f9c0744a591d73bd1c5acc1709071e1be86000c10
91608263a06a0b481fe41fed07828b0e4846fbdde761b612d8e403bb4317f5ec
null
[]
187
2.4
phlo-dagster
0.1.2
Dagster service plugin for Phlo
Dagster service plugin for Phlo.
text/plain
null
Phlo Team <team@phlo.dev>
null
null
MIT
null
[]
[]
null
null
>=3.11
[]
[]
[]
[ "phlo>=0.1.0", "dagster-graphql>=1.12.1", "pyyaml>=6.0.1", "requests>=2.32.5", "phlo-observatory>=0.1.0; extra == \"observatory\"", "pytest>=7.0; extra == \"dev\"", "ruff>=0.1.0; extra == \"dev\"", "phlo-alerting>=0.1.0; extra == \"alerting\"", "phlo-iceberg>=0.1.0; extra == \"iceberg\"" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:16:59.772769
phlo_dagster-0.1.2.tar.gz
48,388
b6/aa/4c4beea98b095fd776871fd4910cf3fee275f0b488ded2bde3f2ee6ee3d3/phlo_dagster-0.1.2.tar.gz
source
sdist
null
false
1f1a91910abedf7b402722e00ecfffec
ecef4b4a382aa9e498582e8574089d4dbbe88e30ec840860ab5b91315266b551
b6aa4c4beea98b095fd776871fd4910cf3fee275f0b488ded2bde3f2ee6ee3d3
null
[]
187
2.4
phlo
0.5.1
Lakehouse platform
<p align="center"> <img src="docs/assets/phlo.png" alt="Phlo" width="400"> </p> <p align="center"> <strong>Modern data lakehouse platform built on Dagster, DLT, Iceberg, Nessie, and dbt.</strong> </p> <p align="center"> <a href="https://github.com/iamgp/phlo/actions/workflows/ci.yml"><img src="https://github.com/iamgp/phlo/actions/workflows/ci.yml/badge.svg" alt="CI"></a> <a href="https://pypi.org/project/phlo/"><img src="https://img.shields.io/pypi/v/phlo" alt="PyPI"></a> <img src="https://img.shields.io/badge/python-3.11+-blue.svg" alt="Python 3.11+"> </p> ## Features - **Decorator-driven development** - Reduce boilerplate by 74% with `@phlo.operations.ingestion` and `@phlo.quality` - **Write-Audit-Publish pattern** - Git-like branching with automatic quality gates and promotion - **Type-safe data quality** - Pandera schemas enforce validation and generate Iceberg tables - **Plugin architecture** - Extensible via service, source, quality, and transformation plugins - **Observatory UI** - Web-based interface for data exploration, lineage, and monitoring - **Production-ready patterns** - Auto-publishing to Postgres, configurable merge strategies, freshness policies - **Modern tooling** - Built on Dagster, DLT, Iceberg, Nessie, dbt, and Trino ## Quick Start ```bash # Install with default services uv pip install phlo[defaults] # Initialize a new project phlo init my-project cd my-project # Start services and run phlo services start phlo materialize --select "dlt_glucose_entries+" ``` ## Documentation Full documentation at [docs/index.md](docs/index.md): - [Installation Guide](docs/getting-started/installation.md) - [Quickstart Guide](docs/getting-started/quickstart.md) - [Core Concepts](docs/getting-started/core-concepts.md) - [Developer Guide](docs/guides/developer-guide.md) - [Plugin Development](docs/guides/plugin-development.md) - [CLI Reference](docs/reference/cli-reference.md) - [Configuration Reference](docs/reference/configuration-reference.md) - [Operations Guide](docs/operations/operations-guide.md) - [Blog Series](docs/blog/README.md) - 13-part deep dive ## Development ```bash # Services phlo services start # Start all services phlo services stop # Stop services phlo services logs -f # View logs # Development uv pip install -e . # Install Phlo ruff check src/ # Lint ruff format src/ # Format basedpyright src/ # Type check phlo test # Run tests ```
text/markdown
null
null
null
null
null
null
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Database", "Topic :: Software Development :: Libraries :: Python Modules" ]
[]
null
null
>=3.11
[]
[]
[]
[ "asyncpg>=0.30.0", "click>=8.0", "dagster-webserver>=1.12.1", "dagster>=1.12.1", "duckdb>=0.9.0", "fastapi>=0.127.0", "httpx>=0.28.1", "jsonschema>=4.23.0", "pandas>=2.3.3", "pandera>=0.26.1", "psycopg2-binary>=2.9.11", "pydantic-settings>=2.11.0", "python-ulid>=2.0.0", "pyyaml>=6.0.1", "requests>=2.32.5", "rich>=13.0", "structlog>=25.5.0", "phlo-dagster>=0.1.0; extra == \"core-services\"", "phlo-minio>=0.1.0; extra == \"core-services\"", "phlo-nessie>=0.1.0; extra == \"core-services\"", "phlo-postgres>=0.1.0; extra == \"core-services\"", "phlo-trino>=0.1.0; extra == \"core-services\"", "phlo-core-plugins>=0.1.0; extra == \"defaults\"", "phlo-dagster>=0.1.0; extra == \"defaults\"", "phlo-dbt>=0.1.0; extra == \"defaults\"", "phlo-dlt>=0.1.0; extra == \"defaults\"", "phlo-iceberg>=0.1.0; extra == \"defaults\"", "phlo-minio>=0.1.0; extra == \"defaults\"", "phlo-nessie>=0.1.0; extra == \"defaults\"", "phlo-postgres>=0.1.0; extra == \"defaults\"", "phlo-trino>=0.1.0; extra == \"defaults\"", "phlo-openmetadata>=0.1.0; extra == \"openmetadata\"" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:16:58.709424
phlo-0.5.1.tar.gz
89,499
f8/63/5b28c5e171d20fa4ad8de9803e1a9379013915d98ba6239634a1bd979e59/phlo-0.5.1.tar.gz
source
sdist
null
false
2a3064e521c786d56e0f0b85d6d820bc
fdef9968df56b88b952319ee58eb0d844d16cc1cea6d1b2d83bcfe42450bdf48
f8635b28c5e171d20fa4ad8de9803e1a9379013915d98ba6239634a1bd979e59
null
[]
192
2.4
fh-matui
0.9.25
material-ui for fasthtml
# fh-matui <!-- WARNING: THIS FILE WAS AUTOGENERATED! DO NOT EDIT! --> ## What is fh-matui? **fh-matui** is a Python library that brings Google’s Material Design to [FastHTML](https://fastht.ml/) applications. It provides a comprehensive set of pre-built UI components that integrate seamlessly with FastHTML’s hypermedia-driven architecture. Built on top of [BeerCSS](https://www.beercss.com/) (a lightweight Material Design 3 CSS framework), fh-matui enables you to create modern, responsive web interfaces entirely in Python — no JavaScript required. ## ✨ Key Features | Feature | Description | |----|----| | 🎨 **Material Design 3** | Modern, beautiful components following Google’s latest design language | | ⚡ **Zero JavaScript** | Build interactive UIs entirely in Python with FastHTML | | 📱 **Responsive** | Mobile-first design with automatic breakpoint handling | | 🌙 **Dark Mode** | Built-in light/dark theme support with 20+ color schemes | | 🧩 **Composable** | Chainable styling APIs inspired by MonsterUI | | 📊 **Data Tables** | Full-featured tables with pagination, search, sorting, and CRUD | | 🔧 **nbdev-powered** | Literate programming with documentation built from notebooks | ## 🏗️ Architecture ┌─────────────────────────────────────────────────────────────┐ │ fh-matui │ ├─────────────────────────────────────────────────────────────┤ │ Foundations │ Core styling utilities, helpers, enums │ │ Core │ Theme system, MatTheme color presets │ │ Components │ Buttons, Cards, Modals, Forms, Tables │ │ App Pages │ Full-page layouts, navigation patterns │ │ Data Tables │ DataTable, DataTableResource for CRUD │ │ Web Pages │ Landing pages, marketing components │ ├─────────────────────────────────────────────────────────────┤ │ BeerCSS │ Material Design 3 CSS framework │ │ FastHTML │ Python hypermedia web framework │ └─────────────────────────────────────────────────────────────┘ ## 🎨 Available Themes fh-matui includes 15+ pre-configured Material Design 3 color themes: | Theme | Preview | Theme | Preview | |------------------------|---------|------------------------|---------| | `MatTheme.red` | 🔴 | `MatTheme.pink` | 🩷 | | `MatTheme.purple` | 🟣 | `MatTheme.deep_purple` | 💜 | | `MatTheme.indigo` | 🔵 | `MatTheme.blue` | 💙 | | `MatTheme.light_blue` | 🩵 | `MatTheme.cyan` | 🌊 | | `MatTheme.teal` | 🩶 | `MatTheme.green` | 💚 | | `MatTheme.light_green` | 🍀 | `MatTheme.lime` | 💛 | | `MatTheme.yellow` | 🌟 | `MatTheme.amber` | 🧡 | | `MatTheme.orange` | 🟠 | `MatTheme.deep_orange` | 🔶 | | `MatTheme.brown` | 🟤 | `MatTheme.grey` | ⚪ | | `MatTheme.blue_grey` | 🔘 | | | **Usage:** ``` python # Choose your theme app, rt = fast_app(hdrs=[MatTheme.deep_purple.headers()]) ``` ## 🚀 Quick Start Here’s a minimal example to get you started: ``` python from fasthtml.common import * from fh_matui.core import MatTheme from fh_matui.components import Button, Card, FormField # Create a themed FastHTML app app, rt = fast_app(hdrs=[MatTheme.indigo.headers()]) @rt('/') def home(): return Div( Card( H3("Welcome to fh-matui!"), P("Build beautiful Material Design apps with Python."), Button("Get Started", cls="primary"), ), cls="padding" ) serve() ``` ## 📦 Installation ``` bash pip install fh-matui ``` ### Dependencies fh-matui automatically includes: - **python-fasthtml** - The core FastHTML framework - **BeerCSS** - Loaded via CDN for Material Design 3 styling ### What This Code Does 1. **`MatTheme.indigo.headers()`** - Loads BeerCSS with the indigo color scheme 2. **[`Card`](https://abhisheksreesaila.github.io/fh-matui/components.html#card)** - Creates a Material Design card component with elevation 3. **[`FormField`](https://abhisheksreesaila.github.io/fh-matui/components.html#formfield)** - Generates a styled input with floating label 4. **`Button`** - Renders a Material Design button with ripple effects ## 📚 Module Reference | Module | Description | Key Components | |----|----|----| | [Foundations](foundations.html) | Base utilities and helper functions | `BeerHeaders`, `display`, styling helpers | | [Core](core.html) | Theme system and styling | `MatTheme`, color presets, theme configuration | | [Components](components.html) | UI component library | `Button`, [`Card`](https://abhisheksreesaila.github.io/fh-matui/components.html#card), [`FormField`](https://abhisheksreesaila.github.io/fh-matui/components.html#formfield), [`FormModal`](https://abhisheksreesaila.github.io/fh-matui/components.html#formmodal), [`Grid`](https://abhisheksreesaila.github.io/fh-matui/components.html#grid) | | [App Pages](app_pages.html) | Application layouts | Navigation, sidebars, full-page layouts | | [Data Tables](05_table.html) | Data management components | [`DataTable`](https://abhisheksreesaila.github.io/fh-matui/datatable.html#datatable), [`DataTableResource`](https://abhisheksreesaila.github.io/fh-matui/datatable.html#datatableresource), CRUD operations | | [Web Pages](web_pages.html) | Marketing/landing pages | Hero sections, feature grids, testimonials | ## 🛠️ Development ### Install in Development Mode ``` bash # Clone the repository git clone https://github.com/user/fh-matui.git cd fh-matui # Install in editable mode pip install -e . # Make changes under nbs/ directory, then compile nbdev_prepare ``` ## 🤝 Why fh-matui? | Challenge | fh-matui Solution | |----|----| | **CSS complexity** | Pre-built Material Design 3 components via BeerCSS | | **JavaScript fatigue** | FastHTML handles interactivity declaratively | | **Component consistency** | Unified API across all components | | **Dark mode support** | Built-in with automatic system preference detection | | **Responsive design** | Mobile-first grid system and responsive utilities | | **Form handling** | [`FormField`](https://abhisheksreesaila.github.io/fh-matui/components.html#formfield), [`FormGrid`](https://abhisheksreesaila.github.io/fh-matui/components.html#formgrid), [`FormModal`](https://abhisheksreesaila.github.io/fh-matui/components.html#formmodal) for rapid form building | | **Data management** | [`DataTable`](https://abhisheksreesaila.github.io/fh-matui/datatable.html#datatable) and [`DataTableResource`](https://abhisheksreesaila.github.io/fh-matui/datatable.html#datatableresource) for CRUD operations | ## 🤖 For LLM Users fh-matui includes **comprehensive documentation bundles** for Large Language Models, enabling AI assistants (like Claude, ChatGPT, or GitHub Copilot) to help you build FastHTML apps with complete knowledge of the component APIs. ### 📥 Download Context File **[📄 llms-ctx.txt](https://raw.githubusercontent.com/abhisheksreesaila/fh-matui/main/llms-ctx.txt)** — Complete API documentation in LLM-optimized format ### 💡 How to Use 1. **Download the context file** from the link above 2. **Attach it to your LLM conversation** (drag & drop or paste contents) 3. **Ask for implementation** using natural language **Example Prompt:** I'm using fh-matui (context attached). Create a dashboard with: - A sidebar navigation with 5 menu items - A DataTable showing products with pagination - A modal form to add/edit products - Use the deep purple theme The LLM will generate production-ready FastHTML code using the exact component APIs from the documentation. ### 🔄 Staying Up to Date The `llms-ctx.txt` file is automatically regenerated with each release to ensure it stays synchronized with the latest API changes. Always download the version matching your installed package version for the most accurate results. > **📌 Note:** The context file is generated from the same literate > programming notebooks that build the library itself, ensuring 100% > accuracy with the actual implementation. ## 📄 License This project is licensed under the Apache 2.0 License - see the [LICENSE](https://github.com/user/fh-matui/blob/main/LICENSE) file for details. ------------------------------------------------------------------------ **Built with ❤️ using FastHTML and nbdev**
text/markdown
abhishek sreesaila
abhishek.sreesaila@gmail.com
null
null
Apache Software License 2.0
nbdev jupyter notebook python
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "Natural Language :: English", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "License :: OSI Approved :: Apache Software License" ]
[]
https://github.com/abhisheksreesaila/fh-matui
null
>=3.9
[]
[]
[]
[ "python-fasthtml", "fastcore", "markdown" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.13.5
2026-02-21T00:16:03.236235
fh_matui-0.9.25.tar.gz
62,339
0d/4d/2fd0c2ea77324c351621ac0065f377b46722ff2b0905fde390718dfa6d45/fh_matui-0.9.25.tar.gz
source
sdist
null
false
335f834cdc4b921471535b6d69ba0309
79575e9c91a95abda2cbf503a619a5cbf5754b6fba83acf1ee4cf7a0a0968e42
0d4d2fd0c2ea77324c351621ac0065f377b46722ff2b0905fde390718dfa6d45
null
[ "LICENSE" ]
226
2.1
AoE2ScenarioParser
0.7.1
This is a project for editing parts of an 'aoe2scenario' file from Age of Empires 2 Definitive Edition
# AoE2ScenarioParser `AoE2ScenarioParser` is a [Python] library that allows you to edit `aoe2scenario` files from **every** version of **Age of Empires 2 Definitive Edition**. [Python]: https://www.python.org/ # Getting Started Installing using `pip`: ```sh pip install AoE2ScenarioParser ``` More documentation about installing etc. can be found below. ## Documentation Documentation for installation, usage, examples, cheatsheets and API docs can be found on **[GitHub Pages]**. [GitHub Pages]: https://ksneijders.github.io/AoE2ScenarioParser/ ## Quick links - **[Installing]** → A quick guide on how to install `AoE2ScenarioParser` - **[Hello World Example]** → Step-by-step guide to get you going - **[Discord Server]** → For questions about `AoE2ScenarioParser`, [Python] or scenarios in general. - **[API Docs]** → Technical documentation for all exposed functions & classes [Installing]: https://ksneijders.github.io/AoE2ScenarioParser/installation/ [Hello World Example]: https://ksneijders.github.io/AoE2ScenarioParser/hello_world/ [Discord Server]: https://discord.gg/DRUtmugXT3 [API Docs]: https://ksneijders.github.io/AoE2ScenarioParser/api_docs/aoe2_scenario/ # Discord If you have any questions regarding `AoE2ScenarioParser`? [Join the discord]! [Join the discord]: https://discord.gg/DRUtmugXT3 # Support **Every Single Scenario Version\* from Age of Empires 2 Definitive Edition is SUPPORTED!** > Support: `1.36` _Version at Release (November 14th, 2019)_ → `1.56` _Current Version (Since: October 14th, 2025)_ Every single version of **Age of Empires 2 Definitive Edition** is supported*! If a new version of **Age of Empires 2 Definitive Edition** just released, it can take a bit for it to be able to be read. Check the [Discord Server] for more up-to-date information if this is the case. If you find a scenario which can be opened by the game itself, but results in an error when using `AoE2ScenarioParser`, please report it as an issue or in the **#bug‑reports** channel in the [Discord Server]. *: All scenario versions are supported, though older structure versions of the same scenario version are not. For more context see [this Discord post](https://discord.com/channels/866955546182942740/877085102201536553/1372708645711777843). To view the full-blown support table previously shown on this page, visit: [support.md]. [support.md]: https://github.com/KSneijders/AoE2ScenarioParser/blob/master/resources/md/support.md # Progress Every related change to the library is documented and can be found in the [CHANGELOG]. [changelog]: https://github.com/KSneijders/AoE2ScenarioParser/blob/dev/CHANGELOG.md ## Features: `AoE2ScenarioParser` allows you to edit **anything** inside a scenario. For general usability "managers" have been created to make working with the files easier. These managers allow you to quickly change aspects of units, triggers, the map, player data and more! Below is a simplified overview of some of the features: | | Inspect | Add | Edit | Remove | |------------|-------------------|-----|------|--------| | Triggers | ✔️ | ✔️ | ✔️ | ✔️ | | Conditions | ✔️ | ✔️ | ✔️ | ✔️ | | Effects | ✔️ | ✔️ | ✔️ | ✔️ | | Units | ✔️ | ✔️ | ✔️ | ✔️ | | Map | n/a *<sup>1</sup> | ✔️ | ✔️ | ✔️ | | Players | n/a *<sup>1</sup> | ✔️* | ✔️ | ✔️* | | Messages | n/a | ✔️ | ✔️ | ✔️ | *: You can disable or enable players like in the in-game editor (min 1, max 8). *<sup>1</sup>: There's no specific inspection function. Though, they can still be printed with clean formatting. # Authors - [Kerwin Sneijders](https://github.com/KSneijders) (Main Author) - [Alian713](https://github.com/Divy1211) (Dataset Wizard) # License MIT License: Please see the [LICENSE file]. [license file]: https://github.com/KSneijders/AoE2ScenarioParser/blob/dev/LICENSE
text/markdown
null
Kerwin Sneijders <ksneijders-dev@hotmail.com>
null
null
null
null
[ "Programming Language :: Python :: 3", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent" ]
[]
null
null
>=3.8
[]
[]
[]
[ "deprecation", "typing_extensions", "ordered-set==4.1.0" ]
[]
[]
[]
[ "Homepage, https://github.com/KSneijders/AoE2ScenarioParser", "Bug Tracker, https://github.com/KSneijders/AoE2ScenarioParser/issues", "Documentation, https://ksneijders.github.io/AoE2ScenarioParser/", "API Docs, https://ksneijders.github.io/AoE2ScenarioParser/api_docs/aoe2_scenario/", "Changelog, https://github.com/KSneijders/AoE2ScenarioParser/blob/master/CHANGELOG.md" ]
twine/6.1.0 CPython/3.13.7
2026-02-21T00:15:48.722307
aoe2scenarioparser-0.7.1.tar.gz
2,069,475
32/63/0c20da24dc1232ff073ab049e23b864fe444305f958df7dd9e50cbc77b24/aoe2scenarioparser-0.7.1.tar.gz
source
sdist
null
false
1070e95057b325308b490b113f75dae3
390006a469fe4b8318f35ac162cfeecc075980b2451a271ca35d990133468180
32630c20da24dc1232ff073ab049e23b864fe444305f958df7dd9e50cbc77b24
null
[]
0
2.1
chaiverse
1.33.6
Chaiverse
[![Chaiverse Banner](https://i.imgur.com/HGlydGb.png)](https://www.chaiverse.com/) [![PyPI version](https://badge.fury.io/py/chaiverse.svg)](https://badge.fury.io/py/chaiverse) [![first-timers-only Friendly](https://img.shields.io/badge/first--timers--only-friendly-blue.svg)](http://www.firsttimersonly.com/) [![Pull Requests Welcome](https://img.shields.io/badge/PRs-welcome-brightgreen.svg?style=flat)](http://makeapullrequest.com) # $1M LLM Prize Hosted By ![Chai Logo](./resources/chai-logo-black.png#gh-light-mode-only) ![Chai Logo](./resources/chai-logo-white.png#gh-dark-mode-only) [Chaiverse](https://www.chaiverse.com) is part of the Chai Prize Competition, accelerating community AGI. It's the world's first open community challenge with real-user evaluations. You models will be directly deployed on the [Chai App](http://tosto.re/chaiapp) where our over 1.4M daily active users will be providing live feedback. Get to top of the leaderboard and share the $1 million cash prize! ## 🚀 Quick Start with Colab 1. Join the [Chaiverse Discord](https://discord.gg/v6dQNmnevt), our bot will greet you and give you a developer key 🥳 2. Submit a model in < 10 minutes with [Chaiverse Jupyter Notebook Quickstart](https://colab.research.google.com/drive/1FyCamT6icUo5Wlt6qqogHbyREHQQkAY8?usp=sharing) 3. Run through our [Chaiverse Prompt Engineering Guide](https://colab.research.google.com/drive/1eMRidYrys3b1mPrhUOJnfAB3Z7tcCNn0?usp=sharing) to submit models with custom prompts 4. Run through our [Chaiverse: Reward Model Guide](https://drive.google.com/file/d/15lWzRoP0RZ7jVxhas_zQaG2OyvqxaxhT/view?usp=sharing) to submit reward models! ❤️ 5. Run through our [Chaiverse: Blend Model Guide](https://colab.research.google.com/drive/1HeslM8jq7H-bbThoujhE6zM9twKPuukH?usp=sharing) to submit blended models! 6. Take a look at our #new-joiners #dataset-sharing and #ai-discussions channels for easter eggs ## 📼 Local Installation 1. Run `pip install chaiverse` 2. Read through examples shown in the "Quick Start with Colab" section ## :keyboard: Chaiverse CLI You can also interact with Chaiverse directly from your terminal using Chaiverse CLI. Read through our [user guide](https://wild-chatter-b52.notion.site/Chaiverse-CLI-User-Guide-2aba61f9db4a4ac78f26420f4e96ba4c?pvs=4) for instructions. ## 🧠 How Does It Work? - The `chaiverse` pip package provides a way to easily submit your language model, all you need to do is ensure it is on HuggingFace 🤗 - We will automatically **Tritonize** your model for fast inference and host it in our internal GPU cluster 🚀 - Once deployed, Chai users on our platform who enter the **arena mode** will be rating your model directly, providing you with both quantatitive and verbal feedback 📈 - Both the public leaderboard and **user feedback** for your model can be directly downloaded via the `chaiverse` package 🧠 - Cash prizes will be allocated according to your position in the leaderboard 💰 [![Chai Pipeline](https://imgur.com/LtMWOAq.png)](https://www.chaiverse.com) ## Resources | | | | ---------------------------------------------------------------------- | ------------------------------------------------------------------------------------------------| | 🤗 [Chai Huggingface](https://huggingface.co/ChaiML) | Tons of models / datasets for you to finetune on! Including past winner solutions | | 📒 [Fine tuning guide](https://huggingface.co/docs/transformers/training) | Guide on language model finetuning | | 💾 [Datasets](https://github.com/mlabonne/llm-datasets) | Curated list of open-sourced datasets to get started with finetuning | | 💖 [Chaiverse Discord](https://discord.gg/v6dQNmnevt) | Our Chaiverse Competition discord | |🚀 [Deepspeed Guide](https://huggingface.co/docs/transformers/main_classes/deepspeed) | Guide for training with Deepspeed (faster training without GPU bottleneck) | |💬 [Example Conversations](https://huggingface.co/datasets/ChaiML/100_example_conversations) | Here you can find 100 example conversations from the Chai Platform | | ⚒️ [Build with us](https://www.chai-research.com/jobs/)| If you think what we are building is cool, join us!| | ❗ [Competition EULA](https://www.chai-research.com/competition-eula.html)| Covers terms of use and competition agreements| <!-- TODO: fix link to competition eula --> [CHAI RESEARCH CORP. COMPETITION END USER LICENSE AGREEMENT (EULA)](https://www.chai-research.com/competition-eula.html)
text/markdown
Chai Research Corp.
hello@chai-research.com
null
null
MIT
null
[]
[]
https://www.chaiverse.com
null
null
[]
[]
[]
[]
[]
[]
[]
[]
twine/6.1.0 CPython/3.8.18
2026-02-21T00:15:35.263844
chaiverse-1.33.6.tar.gz
51,347
e4/6d/c1112856b31799f205d52f7442f27601fefd76f556ecb7c289e03e05a328/chaiverse-1.33.6.tar.gz
source
sdist
null
false
072edc31924c269cf7652b8f615bb95c
c08a3fb84ed937fd3d8a188493abe46a08027dd88424eff4af305f73c5b04a23
e46dc1112856b31799f205d52f7442f27601fefd76f556ecb7c289e03e05a328
null
[]
149
2.4
energizados
0.1.1.dev0
Framework para detección de fraude en consumo energético
<div id='Título-e-imagen-de-portada' /> # Energizados **Framework para detección de fraude en consumo energético mediante Machine Learning.** *** ## 🚀 Modo Framework (NUEVO) Energizados ahora puede usarse como un **framework extensible** con CLI: ```bash # Instalación pip install -e . # Inicializar un nuevo proyecto energizados init mi_proyecto # Ejecutar pipeline completo energizados run --config configs/pipeline.yaml # Validar configuración energizados validate --config configs/pipeline.yaml ``` ### Características del Framework - **ETL multi-fuente con dependencias**: Define múltiples ETLs que se ejecutan en orden topológico - **Clases base extensibles**: Personaliza ETL, selección de features y modelos - **Configuración YAML**: Workflow predefinido sin escribir código - **Modelos pre-construidos**: LightGBM, CatBoost, Neural Networks, LSTM Para más información, ver la [documentación del framework](docs/GETTING_STARTED.md). ### Nota El **modo notebook** (original) continúa funcionando sin cambios. Ver [Notebook Paso a Paso](notebooks/ejecucion_paso_paso.ipynb). *** <div id='insignias' /> <a target="_blank" href="https://colab.research.google.com/github/EL-BID/Energiza2Cod4Dev/blob/master/notebooks/colab_ejecucion_paso_paso.ipynb"> <img src="https://colab.research.google.com/assets/colab-badge.svg" alt="Open In Colab"/> </a> Este proyecto presenta una guía para la detección de robo eléctrico basado en el aprendizaje automático supervisado utilizando una biblioteca de extracción de variables de series temporales junto con algoritmos de boosting y redes neuronales. Ver [Notebook Paso a Paso](https://github.com/EL-BID/Energiza2Cod4Dev/blob/master/notebooks/ejecucion_paso_paso.ipynb) *** <div id='índice' /> ## Índice * [Título](#Título-e-imagen-de-portada) * [Insignias](#insignias) * [Índice](#índice) * [Descripción del proyecto](#descripción-del-proyecto) * [Guía de instalación](#configuracion-ambiente) * [Guía de usuario](#demostracion-del-proyecto) * [Licencia](#licencia) * [Limitación de responsabilidades](#limitación-de-responsabilidades) *** <div id='descripción-del-proyecto' /> ## Descripción del Proyecto “Energizados” es un proyecto construido para mostrar cómo con el uso de aprendizaje automático se puede ayudar a detectar y disminuir las pérdidas no técnicas reduciendo tiempos de regularización e incrementando la precisión de identificación de fraudes. El marco de detección de pérdidas no técnicas “Energizados” se puede dividir en 3 grandes etapas. La etapa de preprocesamiento de datos, la etapa de construcción de modelos simples basados en reglas o modelos baselines, luego el desarrollo de modelos más complejos como los supervisados y finalmente la etapa de evaluación de modelos. <div style="width: 1000px; height: 600px;"> <img src="img/Pryecto-Energiza2_V23.png" width="100%" height="100%"> </div> ### ***Etapa 1 : Preprocesamiento de datos / Exploración / Entendimiento*** Se requiere de esta etapa para tomar los datos en crudos y darles una estructura de datos significativa. Aquí es importante remarcar que un entendimiento y exploración de datos también es primordial ya que en este paso también se puede decidir que otras variables pueden usarse en el proceso de detección de pérdidas no técnicas. En nuestras pruebas, Energizados se evaluó en dos conjuntos de datos proveídos por dos empresas de distribución de energía. Donde se observó que además de las series de consumos mensuales, incorporar variables que describen a los usuarios también ayudan en la detección de usuarios fraudulentos. Entre otras cosas que se pueden observar en esta etapa es la proporción de usuario fraudulentos y no fraudulentos. En este tipo de problemas es común tener clases desbalanceadas, por lo general la proporción de usuarios fraudulentos no supera el 10%. ### ***Etapa 2 : Construcción de modelos*** En esta etapa primeramente se evaluaron modelos simples o modelos baselines para luego desarrollar modelos más complejos. #### ***Modelos simples*** Modelos que a través de reglas analíticas pueden detectar un comportamiento anómalo en el consumo de energía de los usuarios. Estas reglas en general se derivan luego de hacer un análisis exploratorio de los datos y también de conocimientos de expertos. - __Cambio o disminución en el consumo de energía:__ Regla que evalúa si un usuario disminuyó dramáticamente su consumo actual con respecto a periodos anteriores. - __Consumo constante:__ Regla que evalúa si el consumo fue constante por un periodo largo de tiempo. #### ***Modelos supervisados*** En lo que respecta a la construcción de los modelos supervisados, se siguieron los siguientes pasos: - Construcción de variables. - Selección de variables. - Manejo de datos desbalanceados. - Optimización de hiperparámetros. - Entrenamiento final, con los hiperparámetros encontrados en todo el conjunto de datos. ##### ***Ingeniería de variables*** La ingeniería de variables es el proceso de extracción y selección de las variables más importantes de los datos dados, normalmente se realiza para mejorar la capacidad de aprendizaje de los modelos en ML. Esta etapa puede ser dividida en 2 subtareas, una tarea para la extracción o derivación de nuevas variables y otra tarea para la selección de las variables más importantes. _Extracción de variables:_ En lo que respecta a los consumos mensuales de usuarios dados, estos datos por si solo carecen de características estadísticas que reflejen adecuadamente los patrones subyacentes en los datos de consumo de los usuarios, haciendo que los modelos de detección de robo de energía sean menos eficientes. Por lo tanto se crearon nuevas variables derivadas de los consumos mensuales de energía. Estas variables adicionales pueden ser divididas en 3 tipos: - __Estadísticas__: máximo, promedio, mínimo, mediana, son una muestra de los valores estadísticos calculados; - __Espectrales derivadas de la serie de consumo__: distancia de la señal, pendiente de la señal, varianza de la señal, etc.; - __Temporales__: autocorrelación entre las variables, entropía, centroides, entre otros. En lo que respecta a variables que caracterizan a los usuarios, por ejemplo, tipo de tarifa y actividad económica del usuario. De estas se derivaron variables a través de procesamiento clásicos de variables categóricas, entre los cuales se pueden destacar, creación de variables dummy, reducción de cardinalidad y encoding. _Selección de variables:_ La selección de variables es el proceso de identificar un subconjunto representativo de variables de un grupo más grande. En el desarrollo de los modelos, ejecutamos los siguientes pasos: - Eliminación de variables constantes o con muy baja variabilidad. - Eliminación de variables altamente correlacionadas entre sí. - Selección de grupo de variables relevantes para la detección de robo.(Boruta) ##### ***Entrenamiento de modelos*** En cualquier técnica de aprendizaje automático supervisado, los datos etiquetados se proporcionan inicialmente al clasificador de aprendizaje para su propósito de entrenamiento. Luego, el modelo entrenado se evalúa por su capacidad para predecir y generalizar los datos no etiquetados de manera eficiente. Como ocurre en la mayoría de los conjuntos de datos de detección de pérdidas no técnicas, estos están desbalanceados. Los datos desbalanceados generalmente se refieren a una situación que enfrentan los problemas de clasificación donde las clases no están representadas por igual. Actualmente en la literatura se encuentran muchos métodos para poder abordar el problema de datos desbalanceados, además también existen paquetes software que automatizan el proceso y se pueden utilizar con python (Imbalearn). En lo que respecta a lo que se usó en el siguiente marco de trabajo se utilizaron las siguientes estrategias: - Sobremuestreo (oversampling): Consiste en generar nuevas muestras de la clase que está infrarrepresentadas. - Sub–muestreo (under sampling): Consiste en eliminar ejemplos o filas de la clase mayoritaria. Finalmente, se utilizó la opción ‘under-sampling’, ya que se obtuvieron mejores resultados en la fase de desarrollo. Para la optimización de hiperparámetros se siguió la estrategia de búsqueda aleatoria (Random Search), la cual consiste en muestrar valores posibles de los hiperparámetros y quedarse con aquellos que, al ser incluidos en el modelo, hayan generado mejores resultados en las métricas. A continuación se da una breve descripción de los modelos usados. ***Light Gradient Boosting Machine (Light GBM)*** El modelo Light Gradient Boosting Machine (LGBM). LGBM, es un modelo dentro del “gradient boosting framework”, ya que usa algoritmos de aprendizaje basados en árboles de decisión. En particular, LGBM es un modelo de ensamble (combinación de múltiples modelos) dado por el método de boosting. Este método comienza ajustando un modelo inicial en los datos y luego construye un segundo modelo que se enfoca en predecir con precisión los casos en los que el primer modelo tiene un desempeño deficiente. Se espera que la combinación de estos dos modelos sea mejor que cualquiera de los modelos por separado. Esta operación se puede hacer de manera sucesiva, construyendo varios modelos que reduzcan el error del anterior. Algunas características que se destacan en LGBM son la velocidad de entrenamiento (light), la capacidad de manejar grandes volúmenes de datos utilizando poca memoria y el manejo de valores faltantes. ***Red Neuronal*** Las redes neuronales son sistemas de procesamiento de la información cuya estructura y funcionamiento están basados en las redes neuronales biológicas. Estos sistemas están compuestos por elementos simples denominados nodos o neuronas, los cuales se organizan por capas. Cada neurona está conectada con otra a partir de unos enlaces denominados conexiones, cada una de las cuales tiene un peso asociado. Estos pesos son valores numéricos modificables y representan la información que será usada por la red para resolver el problema que se presente. Las conexiones entre las capas tienen un peso o valor asignado, el cual es importante para el aprendizaje de la red. __Multicapa__ La red neuronal multicapa es una red donde todas las señales van en una misma dirección de neurona en neurona, esto se denomina feedforward. En lo que respecta a la problemática de detección de fraude, las entradas son las variables con sus respectivos pre-procesamientos descritos en las secciones anteriores y la capa de salida nos da la probabilidad de que un usuario esté cometiendo fraude. <div> <img src="img/multicapa.png" width="40%" height="40%"> </div> __Concatenación LSTM - Multicapa__ Este es un tipo de red neuronal más compleja que la anterior, en éstas las neuronas presentan conexiones que pueden ir al siguiente nivel de neuronas y a su vez, conexiones que pueden ir al nivel anterior, se denominan conexiones feedforward y feedback. Es una red neuronal que contiene ciclos internos que realimentan la red, generando así memoria. Dicha memoria le permite a la red aprender y generalizar a lo largo de secuencias de entradas en lugar de patrones individuales. Se utilizan en problemas donde la secuencia de los datos sí importa, como por ejemplo en problemas de series de tiempo. En lo que respecta al problema abordado, los consumos de energía mensuales de los usuarios se los puede tratar como una serie de tiempo. En la siguiente figura se observa como se combinó una red lstm con una red multicapa para la detección de fraudes. <div> <img src="img/LSTM.png" width="40%" height="40%"> </div> ### ***Etapa 3 : Evaluación de modelos*** En cualquier técnica de aprendizaje automático supervisado, los datos etiquetados se proporcionan inicialmente al clasificador de aprendizaje para su propósito de entrenamiento. Luego, el modelo entrenado se evalúa por su capacidad para predecir y generalizar los datos no etiquetados de manera eficiente. El rendimiento de dichos modelos se evalúa en función de una serie de métricas de evaluación del rendimiento. En lo que respecta a este proyecto evaluamos en la siguiente métrica. _Auc-roc_: La curva AUC - ROC es una medida de rendimiento para los problemas de clasificación que tiene en cuenta varios ajustes de umbral. La ROC es una curva de probabilidad y la AUC representa el grado o la medida de separabilidad. Indica en qué medida el modelo es capaz de distinguir entre clases. Cuanto más alto sea el AUC, mejor será el modelo para predecir 0s como 0s y 1s como 1s. Por analogía, cuanto mayor sea el AUC, mejor será el modelo para distinguir entre usuarios fraudulentos y aquellos que no lo son. Esta curva tiene en el eje-y la medida TPR (True Positive Rate) y en el eje-x la medida FPR (False Positive Rate): - TPR = Recall = TP / (TP+FN) - FPR = FP / (FP+TN) <div> <img src="img/roc_curve.png" width="20%" height="20%"> </div> *** <div id='configuracion-ambiente' /> ## Guía de instalación El proyecto contiene la siguiente estructura de carpetas: ~~~ Energizado: |--- datos |--- notebooks |--- src : | |--- modeling | |--- preprocessing | |--- helper ~~~ - datos : contiene el conjunto de datos para poder ejecutar el código. - notebooks : contiene las notebooks de ejecución. Existen dos notebooks una para ser ejecutada en Google-Colab y la otra para ejecutar en un entorno local. - src : contiene los módulos python que dan soporte al proyecto. Como se mencionó existen dos notebooks para poder ejecutar el marco de detección de pérdidas no técnicas en la distribución de energía. Si se quiere ejecutar el proyecto en Google-Colab seguir las instrucciones dentro de la notebook para colab. Si se quiere ejecutar en un entorno local, seguir los siguientes pasos: 1. Tener instalada alguna distribución de Python 3.6 o superior. 2. Tener instalado jupyter lab. 3. Descarga/clonar el proyecto de github. 4. Lanzar jupyter lab ``` jupyter lab ``` 5. Posicionarse en la carpeta del proyecto. 6. Abrir un consola de comando en el entorno de jupyter lab. 7. Instalar los requerimientos utilizando ``` pip install -r requirements.txt ``` *** <div id='demostracion-del-proyecto' /> ## Guía de usuario Para hacer un demostración del uso de Energizados, compartimos un conjunto de datos anonimizado. El dataset está conformado de la siguiente manera. - __Cantidad de registros__ : 42500 - __Cantidad de columnas__ : 19 - __% de fraudulentos__ : 5.8 % Descripción de las columnas: | Variable | Descripción | Tipo de dato | Cardinalidad | | :--- | :--- | :--- | :--- | | Consumo de energía mensual | Indica el comportamiento de consumo a nivel mensual de los usuarios. Se consideran los últimos 12 consumos.| Numérica | - | | Actividad | Indica a qué actividad económica se dedica el usuario| Categoría | 284 | | Tipo de Tarifa | Tarifa que tipo de tarifa se le cobra al usuario| Categoría | 47 | | Tensión | Tensión instalada al usuario.| Categoría | 18 | | Material instalacion | Indica tipo de material del medidor instalado| Categoría | 39 | | Zona | Indica la ubicación geográfica a la que pertenece el usuario | Categoría | 38 | | Target | Indica si el consumidor es fraudulento o no | Numérica | 0 - 1 | | Fecha inspección | Indica la fecha en que se inspeccionó al usuario| Fecha | - | Luego pare ver el código del marco de desarrollo funcionando compartimos una notebook donde se puede ir ejecutando paso a paso el proceso. Ver [Notebook Paso a Paso](https://github.com/EL-BID/Energiza2Cod4Dev/blob/master/notebooks/ejecucion_paso_paso.ipynb) *** <div id='licencia' /> ## Licencia El siguiente proyecto ha sido financiado por el BID. Ver la siguiente licencia [LICENCIA](https://github.com/EL-BID/Plantilla-de-repositorio/blob/master/LICENSE.md) *** <div id='limitación-de-responsabilidades' /> ## Limitación de responsabilidades El BID no será responsable, bajo circunstancia alguna, de daño ni indemnización, moral o patrimonial; directo o indirecto; accesorio o especial; o por vía de consecuencia, previsto o imprevisto, que pudiese surgir: i. Bajo cualquier teoría de responsabilidad, ya sea por contrato, infracción de derechos de propiedad intelectual, negligencia o bajo cualquier otra teoría; y/o ii. A raíz del uso de la Herramienta Digital, incluyendo, pero sin limitación de potenciales defectos en la Herramienta Digital, o la pérdida o inexactitud de los datos de cualquier tipo. Lo anterior incluye los gastos o daños asociados a fallas de comunicación y/o fallas de funcionamiento de computadoras, vinculados con la utilización de la Herramienta Digital.
text/markdown
null
Energizados Team <contact@energizados.org>
null
null
MIT
ml, machine-learning, fraud-detection, energy, electricity, non-technical-losses, anomaly-detection
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "Intended Audience :: Science/Research", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Topic :: Scientific/Engineering :: Artificial Intelligence" ]
[]
null
null
>=3.10
[]
[]
[]
[ "boruta==0.4.3", "catboost==1.2.8", "click>=8.0", "imbalanced-learn~=0.12.0", "lightgbm~=4.6.0", "numpy>=1.20.0", "pandas~=1.5.3", "pyarrow~=15.0.0", "pyyaml>=6.0", "rich>=13.0", "scikit-learn~=1.4.2", "scipy~=1.9.0", "tsfel==0.1.4", "tqdm>=4.64.0", "unidecode~=1.4.0", "black>=23.0; extra == \"dev\"", "ipykernel>=6.0; extra == \"dev\"", "ipywidgets>=8.0; extra == \"dev\"", "jupyterlab>=4.0; extra == \"dev\"", "mypy>=1.0; extra == \"dev\"", "pytest>=7.0; extra == \"dev\"", "pytest-cov>=4.0; extra == \"dev\"", "ruff>=0.1.0; extra == \"dev\"", "pre-commit<5.0.0,>=4.5.1; extra == \"dev\"", "matplotlib>=3.5.0; extra == \"viz\"", "seaborn>=0.11.2; extra == \"viz\"" ]
[]
[]
[]
[ "Homepage, https://github.com/energizados/energizados", "Documentation, https://energizados.readthedocs.io", "Repository, https://github.com/energizados/energizados", "Issues, https://github.com/energizados/energizados/issues" ]
twine/6.2.0 CPython/3.10.19
2026-02-21T00:14:46.738529
energizados-0.1.1.dev0.tar.gz
67,448
49/42/e49af08852981c043d422d5374b432f896a1744d1abda3591ae2c5658f40/energizados-0.1.1.dev0.tar.gz
source
sdist
null
false
562aabd5df6ed6ee21b3f1651f51433c
594f89a1fae19e6f8ee8325b06721402a6fcf67eb3382d8998e93f6a37e698b7
4942e49af08852981c043d422d5374b432f896a1744d1abda3591ae2c5658f40
null
[ "LICENSE.md" ]
89
2.4
apache-airflow-providers-fab
1.5.4
Provider package apache-airflow-providers-fab for Apache Airflow
.. Licensed to the Apache Software Foundation (ASF) under one or more contributor license agreements. See the NOTICE file distributed with this work for additional information regarding copyright ownership. The ASF licenses this file to you under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at .. http://www.apache.org/licenses/LICENSE-2.0 .. Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. .. Licensed to the Apache Software Foundation (ASF) under one or more contributor license agreements. See the NOTICE file distributed with this work for additional information regarding copyright ownership. The ASF licenses this file to you under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at .. http://www.apache.org/licenses/LICENSE-2.0 .. Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. .. NOTE! THIS FILE IS AUTOMATICALLY GENERATED AND WILL BE OVERWRITTEN! .. IF YOU WANT TO MODIFY TEMPLATE FOR THIS FILE, YOU SHOULD MODIFY THE TEMPLATE `PROVIDER_README_TEMPLATE.rst.jinja2` IN the `dev/breeze/src/airflow_breeze/templates` DIRECTORY Package ``apache-airflow-providers-fab`` Release: ``1.5.4`` `Flask App Builder <https://flask-appbuilder.readthedocs.io/>`__ Provider package ---------------- This is a provider package for ``fab`` provider. All classes for this provider package are in ``airflow.providers.fab`` python package. You can find package information and changelog for the provider in the `documentation <https://airflow.apache.org/docs/apache-airflow-providers-fab/1.5.4/>`_. Installation ------------ You can install this package on top of an existing Airflow 2 installation (see ``Requirements`` below for the minimum Airflow version supported) via ``pip install apache-airflow-providers-fab`` The package supports the following python versions: 3.10,3.11,3.12 Requirements ------------ ========================================== ================== PIP package Version required ========================================== ================== ``apache-airflow`` ``>=2.11.1`` ``apache-airflow-providers-common-compat`` ``>=1.2.1`` ``flask-login`` ``>=0.6.3`` ``flask-session`` ``>=0.8.0`` ``flask`` ``>=2.2,<3`` ``flask-appbuilder`` ``==4.5.4`` ``google-re2`` ``>=1.0`` ``jmespath`` ``>=0.7.0`` ========================================== ================== Cross provider package dependencies ----------------------------------- Those are dependencies that might be needed in order to use all the features of the package. You need to install the specified provider packages in order to use them. You can install such cross-provider dependencies when installing from PyPI. For example: .. code-block:: bash pip install apache-airflow-providers-fab[common.compat] ================================================================================================================== ================= Dependent package Extra ================================================================================================================== ================= `apache-airflow-providers-common-compat <https://airflow.apache.org/docs/apache-airflow-providers-common-compat>`_ ``common.compat`` ================================================================================================================== ================= The changelog for the provider package can be found in the `changelog <https://airflow.apache.org/docs/apache-airflow-providers-fab/1.5.4/changelog.html>`_.
text/x-rst
null
Apache Software Foundation <dev@airflow.apache.org>
null
Apache Software Foundation <dev@airflow.apache.org>
null
airflow-provider, fab, airflow, integration
[ "Development Status :: 5 - Production/Stable", "Environment :: Console", "Environment :: Web Environment", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "Framework :: Apache Airflow", "Framework :: Apache Airflow :: Provider", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Topic :: System :: Monitoring" ]
[]
null
null
~=3.10
[]
[]
[]
[ "apache-airflow-providers-common-compat>=1.2.1", "apache-airflow>=2.11.1", "flask-appbuilder==4.5.4", "flask-login>=0.6.3", "flask-session>=0.8.0", "flask<3,>=2.2", "google-re2>=1.0", "jmespath>=0.7.0", "kerberos>=1.3.0; extra == \"kerberos\"" ]
[]
[]
[]
[ "Bug Tracker, https://github.com/apache/airflow/issues", "Changelog, https://airflow.apache.org/docs/apache-airflow-providers-fab/1.5.4/changelog.html", "Documentation, https://airflow.apache.org/docs/apache-airflow-providers-fab/1.5.4", "Slack Chat, https://s.apache.org/airflow-slack", "Source Code, https://github.com/apache/airflow", "Twitter, https://x.com/ApacheAirflow", "YouTube, https://www.youtube.com/channel/UCSXwxpWZQ7XZ1WL3wqevChA/" ]
twine/6.2.0 CPython/3.12.11
2026-02-21T00:13:33.874617
apache_airflow_providers_fab-1.5.4.tar.gz
63,052
95/10/325261a48f0888d61e259a1124e6a55ddabbe6acaeb69f77bc746ecba5a8/apache_airflow_providers_fab-1.5.4.tar.gz
source
sdist
null
false
5e77949aa4cbca8e3f31a02ae0d4ac40
3884d1f57f61e9fe13b47e699b58edbe7fe79cc0c10b228516f27271aa1237e3
9510325261a48f0888d61e259a1124e6a55ddabbe6acaeb69f77bc746ecba5a8
null
[]
11,346
2.4
breakpoint-ai
0.1.4
Local-first decision engine for baseline vs candidate LLM output checks.
# BreakPoint AI Prevent bad AI releases before they hit production. ```bash pip install breakpoint-ai ``` You change a model. The output looks fine. But: - Cost jumps +38%. - A phone number slips into the response. - The format breaks your downstream parser. BreakPoint catches it before you deploy. It runs locally. Policy evaluation is deterministic from your saved artifacts. It gives you one clear answer: `ALLOW` · `WARN` · `BLOCK` ## Quick Example ```bash breakpoint evaluate baseline.json candidate.json ``` ```text STATUS: BLOCK Reasons: - Cost increased by 38% (baseline: 1,000 tokens -> candidate: 1,380) - Detected US phone number pattern ``` Ship with confidence. ## Lite First (Default) This is all you need to get started: ```bash breakpoint evaluate baseline.json candidate.json ``` Lite is local, deterministic, and zero-config. Out of the box: - Cost: `WARN` at `+20%`, `BLOCK` at `+40%` - PII: `BLOCK` on first detection (email, phone, credit card) - Drift: `WARN` at `+35%`, `BLOCK` at `+70%` - Empty output: always `BLOCK` **Advanced option:** Need config-driven policies, output contract, latency, presets, or waivers? Use `--mode full` and see `docs/user-guide-full-mode.md`. ## Full Mode (If You Need It) Add `--mode full` when you need config-driven policies, output contract, latency, presets, or waivers. Full details: `docs/user-guide-full-mode.md`. ```bash breakpoint evaluate baseline.json candidate.json --mode full --json --fail-on warn ``` ## CI First (Recommended) ```bash breakpoint evaluate baseline.json candidate.json --json --fail-on warn ``` Why this is the default integration path: - Machine-readable decision payload (`schema_version`, `status`, `reason_codes`, metrics). - Non-zero exit code on risky changes. - Easy to wire into existing CI without additional services. Default policy posture (out of the box, Lite): - Cost: `WARN` at `+20%`, `BLOCK` at `+40%` - PII: `BLOCK` on first detection - Drift: `WARN` at `+35%`, `BLOCK` at `+70%` ### Copy-Paste GitHub Actions Gate Use the template: - `examples/ci/github-actions-breakpoint.yml` Copy it to: - `.github/workflows/breakpoint-gate.yml` What `--fail-on warn` means: - Any `WARN` or `BLOCK` fails the CI step. - Exit behavior remains deterministic: `ALLOW=0`, `WARN=1`, `BLOCK=2`. If you only want to fail on `BLOCK`, change: - `BREAKPOINT_FAIL_ON: warn` to: - `BREAKPOINT_FAIL_ON: block` ## Try In 60 Seconds ```bash pip install -e . make demo ``` What you should see: - Scenario A: `BLOCK` (cost spike) - Scenario B: `BLOCK` (format/contract regression) - Scenario C: `BLOCK` (PII + verbosity drift) - Scenario D: `BLOCK` (small prompt change -> cost blowup) ## Four Realistic Examples Baseline for all examples: - `examples/install_worthy/baseline.json` ### 1) Cost regression after model swap ```bash breakpoint evaluate examples/install_worthy/baseline.json examples/install_worthy/candidate_cost_model_swap.json ``` Expected: `BLOCK` Why it matters: output appears equivalent, but cost increases enough to violate policy. ### 2) Structured-output behavior regression ```bash breakpoint evaluate examples/install_worthy/baseline.json examples/install_worthy/candidate_format_regression.json ``` Expected: `BLOCK` Why it matters: candidate drops expected structure and drifts from baseline behavior. ### 3) PII appears in candidate output ```bash breakpoint evaluate examples/install_worthy/baseline.json examples/install_worthy/candidate_pii_verbosity.json ``` Expected: `BLOCK` Why it matters: candidate introduces PII and adds verbosity drift. ### 4) Small prompt change -> big cost blowup ```bash breakpoint evaluate examples/install_worthy/baseline.json examples/install_worthy/candidate_killer_tradeoff.json ``` Expected: `BLOCK` Why it matters: output still looks workable, but detail-heavy prompt changes plus a model upgrade create large cost and latency increases with output-contract drift. More scenario details: - `docs/install-worthy-examples.md` ## CLI Evaluate two JSON files: ```bash breakpoint evaluate baseline.json candidate.json ``` Evaluate a single combined JSON file: ```bash breakpoint evaluate payload.json ``` JSON output for CI/parsing: ```bash breakpoint evaluate baseline.json candidate.json --json ``` Exit-code gating options: ```bash # fail on WARN or BLOCK breakpoint evaluate baseline.json candidate.json --fail-on warn # fail only on BLOCK breakpoint evaluate baseline.json candidate.json --fail-on block ``` Stable exit codes: - `0` = `ALLOW` - `1` = `WARN` - `2` = `BLOCK` Waivers, config, presets: see `docs/user-guide-full-mode.md`. ## Input Schema Each input JSON is an object with at least: - `output` (string) Optional fields used by policies: - `cost_usd` (number) - `model` (string) - `tokens_total` (number) - `tokens_in` / `tokens_out` (number) - `latency_ms` (number) Combined input format: ```json { "baseline": { "output": "..." }, "candidate": { "output": "..." } } ``` ## Python API ```python from breakpoint import evaluate decision = evaluate( baseline_output="hello", candidate_output="hello there", metadata={"baseline_tokens": 100, "candidate_tokens": 140}, ) print(decision.status) print(decision.reasons) ``` ## Additional Docs - `docs/user-guide.md` - `docs/user-guide-full-mode.md` (Full mode: config, presets, environments, waivers) - `docs/terminal-output-lite-vs-full.md` (Lite vs Full terminal output, same format) - `docs/quickstart-10min.md` - `docs/install-worthy-examples.md` - `docs/baseline-lifecycle.md` - `docs/ci-templates.md` - `docs/value-metrics.md` - `docs/policy-presets.md` - `docs/release-gate-audit.md` ## Contact Suggestions and feedback: [c.holmes.silva@gmail.com](mailto:c.holmes.silva@gmail.com) or [open an issue](https://github.com/cholmess/breakpoint-ai/issues).
text/markdown
null
Christopher Holmes <c.holmes.silva@gmail.com>
null
null
null
null
[]
[]
null
null
<3.13,>=3.10
[]
[]
[]
[ "pytest>=8.0.0; extra == \"dev\"", "sentence-transformers>=2.2.2; extra == \"ml\"", "torch>=2.0.0; extra == \"ml\"" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.10.6
2026-02-21T00:13:25.446217
breakpoint_ai-0.1.4.tar.gz
39,759
4c/c6/28aa9ccdfff53f9b160f0408b3b272849ad73e5fc106d1e57da64f3ab402/breakpoint_ai-0.1.4.tar.gz
source
sdist
null
false
c3b5268ebc2491420deb35a20fa55f97
d8358e7a5d924c81ff587d9c38fe34eba9a1c9098f50a46ce962b0b5b6ff5df2
4cc628aa9ccdfff53f9b160f0408b3b272849ad73e5fc106d1e57da64f3ab402
null
[]
212
2.1
scc-firewall-manager-sdk
1.17.104
Cisco Security Cloud Control Firewall Manager API
Use the documentation to explore the endpoints Security Cloud Control Firewall Manager has to offer
text/markdown
Cisco Security Cloud Control TAC
cdo.tac@cisco.com
null
null
null
OpenAPI, OpenAPI-Generator, Cisco Security Cloud Control Firewall Manager API
[]
[]
null
null
null
[]
[]
[]
[ "pydantic>=2", "python-dateutil", "typing-extensions>=4.7.1", "urllib3<2.1.0,>=1.25.3" ]
[]
[]
[]
[ "Documentation, https://scc-firewall-manager-sdk.readthedocs.io/" ]
twine/6.2.0 CPython/3.11.2
2026-02-21T00:12:55.483001
scc_firewall_manager_sdk-1.17.104-py3-none-any.whl
651,002
e3/e6/91eeff5f05ae7dd2c8e2c4f6c43be732569794074605382531bc3fbeb714/scc_firewall_manager_sdk-1.17.104-py3-none-any.whl
py3
bdist_wheel
null
false
bf39349fb63edbf40122c0c3d75e937e
3c4b96030b51e2541c2175ebe34a61b4d894ae457ee375fd924115ba63bd8756
e3e691eeff5f05ae7dd2c8e2c4f6c43be732569794074605382531bc3fbeb714
null
[]
82
2.4
viewx
0.1.9
Librería de visualización adaptable para HTML, Dashboards y PDFs en Python
# 📦 ViewX — Librería de Visualización Adaptativa para Python **ViewX** es un paquete moderno de Python diseñado para generar **páginas HTML interactivas**, **dashboards dinámicos** y **visualizaciones inteligentes** que se adaptan automáticamente a los objetos agregados por el usuario. Este proyecto ofrece una solución **ligera, intuitiva y escalable**, ideal para crear interfaces visuales llamativas sin depender de frameworks pesados… aunque una parte se encuentra basada en Streamlit mediante dependencias opcionales. --- ## ✨ Características principales - ⚡ **Rápido y minimalista**: cero dependencias pesadas por defecto. - 🧩 **API intuitiva**: crea páginas y dashboards en segundos. - 📐 **Diseño adaptativo**: cada componente se acomoda automáticamente. - 🌐 **Modo HTML**: genera páginas `.html` totalmente autónomas. - 📊 **Modo Dashboard**: plantillas escalables con soporte opcional para Streamlit/Dash. - 🛠️ **Extensible**: añade tus propias plantillas y módulos personalizados. - 🔮 **Visión a futuro**: pensado para expandirse a interfaces inteligentes. --- ## Instalacion ```python pip install viewx ``` ## 🚀 Ejemplo rápido ### Crear una página HTML ```python from viewx.datasets import load_dataset from viewx import HTML # ----------------------------- # DATASET # ----------------------------- df = load_dataset("iris.csv") # ----------------------------- # DASHBOARD # ----------------------------- ( HTML( data=df, title="Reporte Iris — ViewX", template_color=1, num_divs=8, num_cols=4, num_rows=5 ) # ===== VALUE BOXES ===== .add_valuebox( title="Filas", value=len(df), icon="📄", slot_grid=("div1", 1, 1, 1, 1) ) # slot_grid = ("div#", fila_inicial, columna_inicial, alto, ancho) .add_valuebox( title="Prom Sepal Length", value=round(df["sepal_length"].mean(), 2), icon="📏", slot_grid=("div2", 1, 2, 1, 1) ) .add_valuebox( title="Prom Petal Width", value=round(df["petal_width"].mean(), 2), icon="🌸", slot_grid=("div3", 1, 3, 1, 1) ) .add_text( "<h2>Iris Dataset Dashboard</h2><p>Este DashBoard fue desarrollado por Emmanuel Ascendra con Viewx</p>", slot_grid=("div4", 1, 4, 1, 1) ) # ===== PLOTS ===== .add_plot( kind="scatter", x="sepal_length", y="sepal_width", title="Sepal Length vs Width", slot_grid=("div5", 2, 1, 2, 2) ) .add_plot( kind="box", x="species", y="petal_width", title="Petal Width por especie", slot_grid=("div6", 4, 1, 2, 2) ) .add_plot( kind="bar", x="species", y="sepal_length", title="Promedio Sepal Length", slot_grid=("div7", 4, 3, 2, 2) ) # ===== TABLE ===== .add_table( columns="all", slot_grid=("div8", 2, 3, 2, 2) ) .show("demo_viewx.html", port=8001) ) ``` ![DashBoardIris](https://raw.githubusercontent.com/GhostAnalyst30/ViewX/main/images_for_git/DashBoard%20Iris.png ) ### Crear un DashBoard ```python from viewx import DashBoard from viewx.datasets import load_dataset df = load_dataset("iris.csv") db = DashBoard(df, title="StreamOps: Mini Dashboard", title_align="center") db.set_theme(background="#071021", text="#E9F6F2", primary="#19D3A3", card="#0b1620") # Sidebar db.add_sidebar(db.comp_text("Parámetros del reporte")) db.add_sidebar(db.comp_metric("Longitud del dataset", df.shape[0])) db.add_sidebar(db.comp_metric("Cantidad de Flores", df["species"].unique().shape[0])) # Main layout db.add_blank() db.add_row( col_widths=[1, 2, 1], components=[ db.comp_blank(), db.comp_plot(x="sepal_length", y="sepal_width", kind="scatter", color="#FFB86B"), db.comp_metric("sepal_width", df["sepal_width"].sum(), delta="▲ 5%") ] ) db.add_tabs({ "Overview": [ db.comp_title("Resumen por Región"), db.comp_table() ], "Details": [ db.comp_title("Distribución de Flores"), db.comp_plot(x="species", y=None, kind="hist", color="#7C4DFF") ] }) db.add_expander("Detalles técnicos", [ db.comp_text("Este panel fue generado automáticamente."), db.comp_text("Metadata: filas=" + str(len(df)), size="12px") ], expanded=True) db.run(open_browser=True) ``` ![DashBoard Streamlit](https://raw.githubusercontent.com/GhostAnalyst30/ViewX/main/images_for_git/DashBoard_Streamlit_1.png) ### Crear un Reporte ```python from viewx.datasets import load_dataset import seaborn as sns import matplotlib.pyplot as plt # =============================== # 1️⃣ CREAR REPORTE # =============================== r = Report( title="Reporte Técnico ViewX", author="Emmanuel Ascendra" ) # =============================== # 2️⃣ TEXTO # =============================== r.add_text("Este reporte demuestra todas las capacidades del motor ViewX.\n") r.add_text("Texto importante en negrita.", bold=True) # =============================== # 3️⃣ SECCIONES # =============================== with r.doc.create(r.add_section("Introducción")): r.add_text( "ViewX es un motor de generación de reportes científicos " "capaz de producir documentos profesionales usando Python." ) # =============================== # 4️⃣ SUBSECCIÓN # =============================== with r.doc.create(r.add_subsection("Características principales")): r.add_itemize([ "Texto estructurado", "Imágenes", "Tablas", "Código", "Gráficos científicos", "Multicolumnas", "Cajas de información" ]) # =============================== # 5️⃣ TABLA # =============================== with r.doc.create(r.add_section("Tabla de resultados")): r.add_table( headers=["Modelo", "Accuracy", "F1"], rows=[ ["Regresión", 0.82, 0.79], ["Árbol", 0.91, 0.88], ["Red neuronal", 0.94, 0.92], ], caption="Comparación de modelos" ) # =============================== # 6️⃣ IMAGEN # =============================== with r.doc.create(r.add_section("Visualización")): r.add_image( path="assets/ejemplo.png", caption="Imagen de prueba", width="0.6\\linewidth" ) # =============================== # 7️⃣ CÓDIGO # =============================== with r.doc.create(r.add_section("Código de ejemplo")): r.add_code(""" import numpy as np x = np.linspace(0, 10, 50) y = np.sin(x) """) # =============================== # 8️⃣ MULTICOLUMNAS # =============================== with r.doc.create(r.add_section("Análisis en dos columnas")): r.begin_multicols(2) r.add_text( "Este bloque demuestra cómo dividir el contenido " "en múltiples columnas dentro del mismo documento." ) r.add_itemize([ "Ideal para papers", "Mejora lectura", "Ahorra espacio" ]) r.end_multicols() # =============================== # 9️⃣ CAJA DESTACADA # =============================== with r.doc.create(r.add_section("Nota importante")): r.add_box( title="Observación clave", content="Todos los elementos se generan directamente desde Python.", color="green!20" ) # =============================== # 🔟 GRÁFICO SIMPLE # =============================== with r.doc.create(r.add_section("Gráfico simple")): r.add_plot( x=[0, 1, 2, 3, 4], y=[0, 1, 4, 9, 16], caption="Crecimiento cuadrático" ) # =============================== # 1️⃣1️⃣ MULTIGRÁFICO # =============================== with r.doc.create(r.add_section("Gráficos múltiples")): r.add_multiplot( plots=[ ([0, 1, 2, 3], [0, 1, 4, 9]), ([0, 1, 2, 3], [0, 1, 8, 27]), ], caption="Comparación de funciones" ) # =============================== # 1️⃣2️⃣ SALTO DE PÁGINA # =============================== r.new_page() r.add_text("Contenido en una nueva página.") # =============================== # 1️⃣3️⃣ GENERAR PDF # =============================== r.build("reporte_demo") ``` ![Report PDF](https://raw.githubusercontent.com/GhostAnalyst30/ViewX/main/images_for_git/Report_pdf_1.png) ## 🤝 Contribuciones ¡Todas las ideas, mejoras y plantillas son bienvenidas! ViewX está diseñado para crecer y evolucionar con la comunidad. ## 📬 Contacto: ascendraemmanuel@gmail.com
text/markdown
Emmanuel Ascendra Perez
ascendraemmanuel@gmail.com
null
null
null
null
[ "Development Status :: 3 - Alpha", "Intended Audience :: Science/Research", "Intended Audience :: Developers", "Topic :: Scientific/Engineering :: Visualization", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13" ]
[]
https://github.com/GhostAnalyst30/ViewX
null
>=3.8
[]
[]
[]
[ "numpy>=1.25.0", "pandas>=2.1.0", "matplotlib>=3.8.0", "pylatex>=1.4.2", "seaborn>=0.12.2", "plotly>=6.0.0", "streamlit>=1.32.0", "statslibx>=0.2.2", "streamlit>=1.32.0; extra == \"streamlit\"", "dash>=2.14.0; extra == \"dash\"", "seaborn>=0.12.2; extra == \"viz\"", "plotly>=6.0.0; extra == \"viz\"", "pylatex>=1.5.0; extra == \"pdf\"", "streamlit>=1.32.0; extra == \"all\"", "dash>=2.14.0; extra == \"all\"", "seaborn>=0.12.2; extra == \"all\"", "plotly>=6.0.0; extra == \"all\"", "pylatex>=1.5.0; extra == \"all\"" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.13.7
2026-02-21T00:10:26.209627
viewx-0.1.9.tar.gz
4,671,042
2d/7d/007971ae67940ada5824d1e0e2c7bbcdb23748083f487ef3d869e879a2c9/viewx-0.1.9.tar.gz
source
sdist
null
false
177017d91443962444255af16457206c
320e17a3fe997defd571fb4c0d54d59ca2f28c9bc12f521c3c18407240fc74f5
2d7d007971ae67940ada5824d1e0e2c7bbcdb23748083f487ef3d869e879a2c9
null
[]
199
2.4
xlm-core
0.1.4a0
XLM Framework
XLM is a unified framework for developing and comparing small non-autoregressive language models. It uses PyTorch as the deep learning framework, PyTorch Lightning for training utilities, and Hydra for configuration management. XLM provides core components for flexible data handling and training, useful architectural implementations for non-autoregressive workflows, and support for arbitrary runtime code injection. Custom model implementations that leverage the core components of xlm can be found in the xlm-models package. The package also includes a few preconfigured synthetic planning and language-modeling datasets. Usage: pip install xlm-core Command usage: xlm job_type=[JOB_TYPE] job_name=[JOB_NAME] experiment=[CONFIG_PATH] The job_type argument can be one of train ,eval and generate. The experiment argument should point to the root hydra config file.
null
Dhruvesh Patel, Benjamin Rozonoyer, Sai Sreenivas Chintha, Durga Prasad Maram
null
null
null
null
AI, ML, Machine Learning, Deep Learning, Non-Autoregressive Language Models
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Science/Research", "Programming Language :: Python :: 3.11", "Topic :: Scientific/Engineering :: Artificial Intelligence" ]
[]
null
null
>=3.11
[]
[]
[]
[ "lightning==2.5.2", "fsspec", "torch", "jupyter", "hydra-core", "hydra-colorlog", "hydra-joblib-launcher", "hydra-submitit-launcher", "datasets<4.0.0,>=3.3.2", "wandb", "transformers", "more-itertools", "torchdata>=0.9.0", "rich", "python-dotenv", "jaxtyping", "tensorboard", "torch-ema", "pydot", "tabulate", "pandas", "simple_slurm" ]
[]
[]
[]
[ "Source Code, https://github.com/dhruvdcoder/xlm-core" ]
twine/6.2.0 CPython/3.14.2
2026-02-21T00:08:48.787222
xlm_core-0.1.4a0.tar.gz
140,327
ac/96/19707e6325c5374dc55ed4db49fe075cac492caa36a31761a16dce00027c/xlm_core-0.1.4a0.tar.gz
source
sdist
null
false
f4322ca1ccdd2d0528e8943773be02e8
9ee9e5960456ac66c24501af3d27057e16f737cf1e5d27184270c1c2c30ee2f9
ac9619707e6325c5374dc55ed4db49fe075cac492caa36a31761a16dce00027c
null
[]
188
2.4
nvflare
2.7.2rc10
Federated Learning Application Runtime Environment
<img src="https://raw.githubusercontent.com/NVIDIA/NVFlare/main/docs/resources/nvidia_eye.wwPt122j.png" alt="NVIDIA Logo" width="200"> # NVIDIA FLARE [Website](https://nvidia.github.io/NVFlare) | [Paper](https://arxiv.org/abs/2210.13291) | [Blogs](https://developer.nvidia.com/blog/tag/federated-learning) | [Talks & Papers](https://nvflare.readthedocs.io/en/main/publications_and_talks.html) | [Research](./research/README.md) | [Documentation](https://nvflare.readthedocs.io/en/main) [![Blossom-CI](https://github.com/NVIDIA/nvflare/workflows/Blossom-CI/badge.svg?branch=main)](https://github.com/NVIDIA/nvflare/actions) [![documentation](https://readthedocs.org/projects/nvflare/badge/?version=main)](https://nvflare.readthedocs.io/en/main/?badge=main) [![license](https://img.shields.io/badge/License-Apache%202.0-brightgreen.svg)](./LICENSE) [![pypi](https://badge.fury.io/py/nvflare.svg)](https://badge.fury.io/py/nvflare) [![pyversion](https://img.shields.io/pypi/pyversions/nvflare.svg)](https://badge.fury.io/py/nvflare) [![downloads](https://static.pepy.tech/badge/nvflare)](https://pepy.tech/project/nvflare) [![Ask DeepWiki](https://deepwiki.com/badge.svg)](https://deepwiki.com/NVIDIA/NVFlare) [NVIDIA FLARE](https://nvidia.github.io/NVFlare/) (**NV**IDIA **F**ederated **L**earning **A**pplication **R**untime **E**nvironment) is a domain-agnostic, open-source, extensible Python SDK that allows researchers and data scientists to adapt existing ML/DL workflows to a federated paradigm. It enables platform developers to build a secure, privacy-preserving offering for a distributed multi-party collaboration. ## Features FLARE is built on a componentized architecture that allows you to take federated learning workloads from research and simulation to real-world production deployment. Application Features * Support both deep learning and traditional machine learning algorithms (eg. PyTorch, TensorFlow, Scikit-learn, XGBoost etc.) * Support horizontal and vertical federated learning * Built-in Federated Learning algorithms (e.g., FedAvg, FedProx, FedOpt, Scaffold, Ditto, etc.) * Support multiple server and client-controlled training workflows (e.g., scatter & gather, cyclic) and validation workflows (global model evaluation, cross-site validation) * Support both data analytics (federated statistics) and machine learning lifecycle management * Privacy preservation with differential privacy, homomorphic encryption, private set intersection (PSI) From Simulation to Real-World * FLARE Client API to transition seamlessly from ML/DL to FL with minimal code changes * Simulator and POC mode for rapid development and prototyping * Fully customizable and extensible components with modular design * Deployment on cloud and on-premise * Dashboard for project management and deployment * Security enforcement through federated authorization and privacy policy * Built-in support for system resiliency and fault tolerance > _Take a look at [NVIDIA FLARE Overview](https://nvflare.readthedocs.io/en/main/flare_overview.html) for a complete overview, and [What's New](https://nvflare.readthedocs.io/en/main/whats_new.html) for the lastest changes._ ## Installation To install the [current release](https://pypi.org/project/nvflare/): ``` $ python -m pip install nvflare ``` For detailed installation please refer to [NVIDIA FLARE installation](https://nvflare.readthedocs.io/en/main/installation.html). ## Getting Started * To get started, refer to [getting started](https://nvflare.readthedocs.io/en/main/getting_started.html) documentation * Structured, self-paced learning is available through curated tutorials and training paths on the website. * DLI courses: * https://learn.nvidia.com/courses/course-detail?course_id=course-v1:DLI+S-FX-28+V1 * https://learn.nvidia.com/courses/course-detail?course_id=course-v1:DLI+S-FX-29+V1 * visit developer portal https://developer.nvidia.com/flare ## Community We welcome community contributions! Please refer to the [contributing guidelines](./CONTRIBUTING.md) for more details. Ask and answer questions, share ideas, and engage with other community members at [NVFlare Discussions](https://github.com/NVIDIA/NVFlare/discussions). ## Related Talks and Publications Take a look at our growing list of [talks and publications](https://nvflare.readthedocs.io/en/main/publications_and_talks.html), and [technical blogs](https://developer.nvidia.com/blog/tag/federated-learning) related to NVIDIA FLARE. ## License NVIDIA FLARE is released under an [Apache 2.0 license](./LICENSE).
text/markdown; charset=UTF-8
null
null
null
null
null
null
[ "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "License :: OSI Approved :: Apache Software License", "Operating System :: POSIX :: Linux" ]
[]
https://github.com/NVIDIA/NVFlare
null
>=3.9
[]
[]
[]
[ "cryptography>=36.0.0", "Flask==3.0.2", "Flask-JWT-Extended==4.6.0", "Flask-SQLAlchemy==3.1.1", "grpcio>=1.62.1", "gunicorn>=22.0.0", "numpy", "protobuf>=4.24.4", "psutil>=5.9.1", "PyYAML>=6.0", "requests>=2.28.0", "msgpack>=1.0.3", "docker>=6.0", "aiohttp", "pyhocon", "pydantic>=2.0", "safetensors", "tenseal==0.3.15; extra == \"he\"", "openmined.psi==2.0.5; extra == \"psi\"", "torch; extra == \"pt\"", "torchvision; extra == \"pt\"", "scikit-learn; extra == \"sklearn\"", "pandas>=1.5.1; extra == \"sklearn\"", "mlflow; extra == \"tracking\"", "wandb; extra == \"tracking\"", "tensorboard; extra == \"tracking\"", "datadog; extra == \"monitoring\"", "omegaconf; extra == \"config\"", "tenseal==0.3.15; extra == \"app-opt\"", "openmined.psi==2.0.5; extra == \"app-opt\"", "torch; extra == \"app-opt\"", "torchvision; extra == \"app-opt\"", "scikit-learn; extra == \"app-opt\"", "pandas>=1.5.1; extra == \"app-opt\"", "mlflow; extra == \"app-opt\"", "wandb; extra == \"app-opt\"", "tensorboard; extra == \"app-opt\"", "datadog; extra == \"app-opt\"", "pytorch_lightning; extra == \"app-opt\"", "xgboost; extra == \"app-opt\"", "bitsandbytes; extra == \"app-opt\"", "torch; extra == \"app-opt-mac\"", "torchvision; extra == \"app-opt-mac\"", "scikit-learn; extra == \"app-opt-mac\"", "pandas>=1.5.1; extra == \"app-opt-mac\"", "mlflow; extra == \"app-opt-mac\"", "wandb; extra == \"app-opt-mac\"", "tensorboard; extra == \"app-opt-mac\"", "omegaconf; extra == \"core-opt\"", "sphinx>=4.1.1; extra == \"doc\"", "sphinx_rtd_theme; extra == \"doc\"", "recommonmark; extra == \"doc\"", "sphinx-copybutton; extra == \"doc\"", "sphinxcontrib-jquery; extra == \"doc\"", "omegaconf; extra == \"all\"", "tenseal==0.3.15; extra == \"all\"", "openmined.psi==2.0.5; extra == \"all\"", "torch; extra == \"all\"", "torchvision; extra == \"all\"", "scikit-learn; extra == \"all\"", "pandas>=1.5.1; extra == \"all\"", "mlflow; extra == \"all\"", "wandb; extra == \"all\"", "tensorboard; extra == \"all\"", "datadog; extra == \"all\"", "pytorch_lightning; extra == \"all\"", "xgboost; extra == \"all\"", "bitsandbytes; extra == \"all\"", "omegaconf; extra == \"all-mac\"", "torch; extra == \"all-mac\"", "torchvision; extra == \"all-mac\"", "scikit-learn; extra == \"all-mac\"", "pandas>=1.5.1; extra == \"all-mac\"", "mlflow; extra == \"all-mac\"", "wandb; extra == \"all-mac\"", "tensorboard; extra == \"all-mac\"", "isort==5.13.2; extra == \"test-support\"", "flake8==7.1.1; extra == \"test-support\"", "black==24.8.0; extra == \"test-support\"", "click==8.1.7; extra == \"test-support\"", "pytest-xdist==3.6.1; extra == \"test-support\"", "pytest-cov==5.0.0; extra == \"test-support\"", "pandas>=1.5.1; extra == \"test-support\"", "nbformat; extra == \"test-support\"", "nbmake; extra == \"test-support\"", "kagglehub; extra == \"test-support\"", "omegaconf; extra == \"test\"", "tenseal==0.3.15; extra == \"test\"", "openmined.psi==2.0.5; extra == \"test\"", "torch; extra == \"test\"", "torchvision; extra == \"test\"", "scikit-learn; extra == \"test\"", "pandas>=1.5.1; extra == \"test\"", "mlflow; extra == \"test\"", "wandb; extra == \"test\"", "tensorboard; extra == \"test\"", "datadog; extra == \"test\"", "pytorch_lightning; extra == \"test\"", "xgboost; extra == \"test\"", "bitsandbytes; extra == \"test\"", "isort==5.13.2; extra == \"test\"", "flake8==7.1.1; extra == \"test\"", "black==24.8.0; extra == \"test\"", "click==8.1.7; extra == \"test\"", "pytest-xdist==3.6.1; extra == \"test\"", "pytest-cov==5.0.0; extra == \"test\"", "pandas>=1.5.1; extra == \"test\"", "nbformat; extra == \"test\"", "nbmake; extra == \"test\"", "kagglehub; extra == \"test\"", "omegaconf; extra == \"test-mac\"", "torch; extra == \"test-mac\"", "torchvision; extra == \"test-mac\"", "scikit-learn; extra == \"test-mac\"", "pandas>=1.5.1; extra == \"test-mac\"", "mlflow; extra == \"test-mac\"", "wandb; extra == \"test-mac\"", "tensorboard; extra == \"test-mac\"", "isort==5.13.2; extra == \"test-mac\"", "flake8==7.1.1; extra == \"test-mac\"", "black==24.8.0; extra == \"test-mac\"", "click==8.1.7; extra == \"test-mac\"", "pytest-xdist==3.6.1; extra == \"test-mac\"", "pytest-cov==5.0.0; extra == \"test-mac\"", "pandas>=1.5.1; extra == \"test-mac\"", "nbformat; extra == \"test-mac\"", "nbmake; extra == \"test-mac\"", "kagglehub; extra == \"test-mac\"", "sphinx>=4.1.1; extra == \"dev\"", "sphinx_rtd_theme; extra == \"dev\"", "recommonmark; extra == \"dev\"", "sphinx-copybutton; extra == \"dev\"", "sphinxcontrib-jquery; extra == \"dev\"", "omegaconf; extra == \"dev\"", "tenseal==0.3.15; extra == \"dev\"", "openmined.psi==2.0.5; extra == \"dev\"", "torch; extra == \"dev\"", "torchvision; extra == \"dev\"", "scikit-learn; extra == \"dev\"", "pandas>=1.5.1; extra == \"dev\"", "mlflow; extra == \"dev\"", "wandb; extra == \"dev\"", "tensorboard; extra == \"dev\"", "datadog; extra == \"dev\"", "pytorch_lightning; extra == \"dev\"", "xgboost; extra == \"dev\"", "bitsandbytes; extra == \"dev\"", "isort==5.13.2; extra == \"dev\"", "flake8==7.1.1; extra == \"dev\"", "black==24.8.0; extra == \"dev\"", "click==8.1.7; extra == \"dev\"", "pytest-xdist==3.6.1; extra == \"dev\"", "pytest-cov==5.0.0; extra == \"dev\"", "pandas>=1.5.1; extra == \"dev\"", "nbformat; extra == \"dev\"", "nbmake; extra == \"dev\"", "kagglehub; extra == \"dev\"", "sphinx>=4.1.1; extra == \"dev-mac\"", "sphinx_rtd_theme; extra == \"dev-mac\"", "recommonmark; extra == \"dev-mac\"", "sphinx-copybutton; extra == \"dev-mac\"", "sphinxcontrib-jquery; extra == \"dev-mac\"", "omegaconf; extra == \"dev-mac\"", "torch; extra == \"dev-mac\"", "torchvision; extra == \"dev-mac\"", "scikit-learn; extra == \"dev-mac\"", "pandas>=1.5.1; extra == \"dev-mac\"", "mlflow; extra == \"dev-mac\"", "wandb; extra == \"dev-mac\"", "tensorboard; extra == \"dev-mac\"", "isort==5.13.2; extra == \"dev-mac\"", "flake8==7.1.1; extra == \"dev-mac\"", "black==24.8.0; extra == \"dev-mac\"", "click==8.1.7; extra == \"dev-mac\"", "pytest-xdist==3.6.1; extra == \"dev-mac\"", "pytest-cov==5.0.0; extra == \"dev-mac\"", "pandas>=1.5.1; extra == \"dev-mac\"", "nbformat; extra == \"dev-mac\"", "nbmake; extra == \"dev-mac\"", "kagglehub; extra == \"dev-mac\"" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.12.1
2026-02-21T00:08:39.790579
nvflare-2.7.2rc10-py3-none-any.whl
2,942,126
9d/b6/d7f9624af6cd51e24fa118447ecc011ec593b87254e968070a80bfd8e384/nvflare-2.7.2rc10-py3-none-any.whl
py3
bdist_wheel
null
false
5ab23595ec0014b9e3fa848c4420ee1f
203ae0d56f0d783900fd3e76b49fedafd10bc95168d23395cd4f6d867aa92708
9db6d7f9624af6cd51e24fa118447ecc011ec593b87254e968070a80bfd8e384
null
[ "LICENSE" ]
82
2.4
pyolive-ng
1.1.4
Python API for Athena pywok (Next Generation with Athena Broker).
pyolive-ng ========== Python API for Athena pywok (Next Generation with Athena Broker). ## 시스템 요구사항 **중요**: `pyolive-ng`는 NNG (Nanomsg Next Gen) v1.11 네이티브 라이브러리가 시스템에 설치되어 있어야 합니다. ### NNG v1.11 설치 #### Linux ```bash # Debian/Ubuntu sudo apt-get install libnng-dev # 또는 소스에서 빌드 (v1.11) wget https://github.com/nanomsg/nng/archive/v1.11.0.tar.gz tar -xzf v1.11.0.tar.gz cd nng-1.11.0 mkdir build && cd build cmake .. make sudo make install ``` #### macOS ```bash # Homebrew brew install nng # 또는 소스에서 빌드 brew install cmake git clone https://github.com/nanomsg/nng.git cd nng git checkout v1.11.0 mkdir build && cd build cmake .. make sudo make install ``` #### Windows - NNG v1.11 DLL을 다운로드하여 시스템 PATH에 추가하거나 - 소스에서 빌드 ### Python 버전 요구사항 **중요**: `pyolive-ng`는 **Python 3.12 이상**이 필요합니다. ```bash # Python 버전 확인 python3 --version # Python 3.12.x 이상이어야 함 ``` ### Python 패키지 설치 Install and update using `pip`: ```bash python3 -m pip install -U pyolive-ng ``` **참고**: - `pynng`는 Python 바인딩일 뿐이며, 실제로는 시스템에 설치된 NNG 네이티브 라이브러리를 사용합니다. - Python 3.12 이상에서 최신 성능 개선 및 기능을 활용할 수 있습니다. ## 설정 파일 `pyolive-ng`는 `$ATHENA_HOME/config/` 디렉토리에서 설정 파일을 읽습니다. ### athena-agent.yaml Athena Broker 연결 설정을 포함합니다: ```yaml broker: hosts: - localhost # 또는 실제 broker 호스트 주소 port: 2736 # Athena Broker 기본 포트 username: guest password: guest worker: reload-on-change: false log: level: INFO rotate: 3 size: 10mb ``` **중요**: - `broker/hosts`는 반드시 설정되어야 합니다. - Athena Broker 서버가 실행 중이어야 합니다. - 연결 실패 시 에러 메시지에 설정 파일 경로와 broker URL이 표시됩니다. ## A Simple Example ### main.py ```python import asyncio import argparse from pyolive.agent import Athena from registry import register_apps async def main(): parser = argparse.ArgumentParser(description='Start Athena pywok') parser.add_argument('-ns', required=True, help='namespace (e.g. dps.msm)') parser.add_argument('-alias', required=True, help='worker alias (e.g. worker)') args = parser.parse_args() agent = Athena(namespace=args.ns, alias=args.alias) register_apps(agent) await agent.run() if __name__ == '__main__': try: asyncio.run(main()) except KeyboardInterrupt: print("\nAthena pywok stopped by user.") ``` ### registry.py ```python from helloworld import HelloWorld def register_apps(agent): app_map = { 'ovm_hello_py': HelloWorld } for app_name, cls in app_map.items(): agent.add_resource(cls, app_name) ``` ### helloworld.py ```python import os from pyolive.status import AppStatus class HelloWorld: def __init__(self, *, logger, channel, context): self.logger = logger self.channel = channel self.context = context async def app_main(self): tag = self.context.get_param('tag') self.logger.info("parameter tag = %s", tag) files = self.context.get_fileset() if files: self.logger.info("%s, filenames = %s", self.context.action_app, files) data = self.context.get_msgbox() if data: self.logger.info("%s, msgbox = %s", self.context.action_app, data) # Example of sending job status message await self.channel.publish_notify(self.context, text="HelloWorld completed") return AppStatus.OK # Developer test mode (optional) if __name__ == '__main__': import asyncio import sys import traceback from pyolive.develop import develop_mode async def run(): try: # user set params = {'tag': 'debug'} infile = "." log, ch, ctx = await develop_mode(infile, params) app = HelloWorld(logger=log, channel=ch, context=ctx) result = await app.app_main() return 0 if result == AppStatus.OK else 1 except Exception as e: print(f"[ERROR] test run failed: {e}", file=sys.stderr) traceback.print_exc() return 2 sys.exit(asyncio.run(run())) ```
text/markdown
Kiro Lee
kiroly@ingris.com
null
null
Proprietary
null
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13" ]
[]
http://www.ingris.com
null
>=3.9
[]
[]
[]
[ "pynng>=0.8.0", "PyYAML>=6.0.1", "watchdog>=6.0.0", "aiomysql>=0.2.0", "asyncpg>=0.30.0" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.14.3
2026-02-21T00:08:02.204631
pyolive_ng-1.1.4-py3-none-any.whl
42,832
b3/1b/2d718d5b496f1a1900a49cc87dd8da316ba176db10fd2fef9ed5404b67c9/pyolive_ng-1.1.4-py3-none-any.whl
py3
bdist_wheel
null
false
9573f41084a6968a0912c99ce7cc18cb
a8292a38ba79814a959d30c76bdcf2261779839e24a6c28c71bd9c459466213f
b31b2d718d5b496f1a1900a49cc87dd8da316ba176db10fd2fef9ed5404b67c9
null
[ "LICENSE.txt" ]
83