metadata_version
string
name
string
version
string
summary
string
description
string
description_content_type
string
author
string
author_email
string
maintainer
string
maintainer_email
string
license
string
keywords
string
classifiers
list
platform
list
home_page
string
download_url
string
requires_python
string
requires
list
provides
list
obsoletes
list
requires_dist
list
provides_dist
list
obsoletes_dist
list
requires_external
list
project_urls
list
uploaded_via
string
upload_time
timestamp[us]
filename
string
size
int64
path
string
python_version
string
packagetype
string
comment_text
string
has_signature
bool
md5_digest
string
sha256_digest
string
blake2_256_digest
string
license_expression
string
license_files
list
recent_7d_downloads
int64
2.4
neptune-query
1.12.0b1
Neptune Query is a Python library for retrieving data from Neptune.
# Neptune Query API The `neptune_query` package is a read-only API for fetching metadata. With the Query API, you can: - List experiments, runs, and attributes of a project. - Fetch experiment or run metadata as a data frame. - Define filters to fetch experiments, runs, and attributes that meet certain criteria. ## Installation ```bash pip install "neptune-query<2.0.0" ``` Set your Neptune API token and project name as environment variables: ```bash export NEPTUNE_API_TOKEN="ApiTokenFromYourNeptuneProfile" ``` ```bash export NEPTUNE_PROJECT="workspace-name/project-name" ``` > **Note:** You can also pass the project path to the `project` argument of any querying function. ## Usage ```python import neptune_query as nq ``` Available functions: - `download_files()` &ndash; download files from the specified experiments. - `fetch_experiments_table()` &ndash; runs as rows and attributes as columns. - (experimental) `fetch_experiments_table_global()` &ndash; like `fetch_experiments_table()`, but searches across all projects that the user has access to. - (experimental) `fetch_metric_buckets()` &ndash; get summary values split by X-axis buckets. - `fetch_metrics()` &ndash; series of float or int values, with steps as rows. - ([runs API](#runs-api)) `fetch_runs_table()` &ndash; like `fetch_experiments_table()`, but for individual runs. - `fetch_series()` &ndash; for series of strings or histograms. - `list_attributes()` &ndash; all logged attributes of the target project's experiment runs. - `list_experiments()` &ndash; names of experiments in the target project. - `set_api_token()` &ndash; set the Neptune API token to use for the session. For details, see the [API reference](./docs/api_reference/). ### Runs API You can target individual runs by ID instead of experiment runs by name. To use the corresponding methods for runs, import the `runs` module: ```python import neptune_query.runs as nq_runs nq_runs.fetch_metrics(...) ``` You can use these methods to target individual runs by ID instead of experiment runs by name. ## Documentation For how-tos and the complete API reference, see the [docs](./docs) directory. ## Examples The following are some examples of how to use the Query API. For all functions and options, see the [API reference](./docs/api_reference/). ### Example 1: Fetch metric values To fetch values at each step, use `fetch_metrics()`. - To filter experiments to return, use the `experiments` parameter. - To specify attributes to include as columns, use the `attributes` parameter. - To limit the returned values, use the available parameters. ```python nq.fetch_metrics( experiments=["exp_dczjz"], attributes=r"metrics/val_.+_estimated$", tail_limit=10, ) ``` ```pycon metrics/val_accuracy_estimated metrics/val_loss_estimated experiment step exp_dczjz 1.0 0.432187 0.823375 2.0 0.649685 0.971732 3.0 0.760142 0.154741 4.0 0.719508 0.504652 ``` ### Example 2: Fetch metadata as one row per run To fetch experiment metadata from your project, use the `fetch_experiments_table()` function. The output mimics the runs table in the web app. - To filter experiments to return, use the `experiments` parameter. - To specify attributes to include as columns, use the `attributes` parameter. ```python nq.fetch_experiments_table( experiments=r"^exp_", attributes=["metrics/train_accuracy", "metrics/train_loss", "learning_rate"], ) ``` ```pycon metrics/train_accuracy metrics/train_loss learning_rate experiment exp_ergwq 0.278149 0.336344 0.01 exp_qgguv 0.160260 0.790268 0.02 exp_hstrj 0.365521 0.459901 0.01 ``` > For series attributes, the value of the last logged step is returned. ### Example 3: Define filters List my experiments that have a "dataset_version" attribute and "val/loss" less than 0.1: ```py from neptune_query.filters import Filter owned_by_me = Filter.eq("sys/owner", "sigurd") dataset_check = Filter.exists("dataset_version") loss_filter = Filter.lt("val/loss", 0.1) interesting = owned_by_me & dataset_check & loss_filter nq.list_experiments(experiments=interesting) ``` ```pycon ['exp_ergwq', 'exp_qgguv', 'exp_hstrj'] ``` Then fetch configs from the experiments, including also the interesting metric: ```py nq.fetch_experiments_table( experiments=interesting, attributes=r"config/ | val/loss", ) ``` ```pycon config/optimizer config/batch_size config/learning_rate val/loss experiment exp_ergwq Adam 32 0.001 0.0901 exp_qgguv Adadelta 32 0.002 0.0876 exp_hstrj Adadelta 64 0.001 0.0891 ``` ### Example 4: Exclude archived runs To exclude archived experiments or runs from the results, create a filter on the `sys/archived` attribute: ```py import neptune_query as nq from neptune_query.filters import Filter exclude_archived = Filter.eq("sys/archived", False) nq.fetch_experiments_table(experiments=exclude_archived) ``` To use this filter in combination with other criteria, use the `&` operator to join multiple filters: ```py name_matches_regex = Filter.name(r"^exp_") exclude_archived = Filter.eq("sys/archived", False) nq.fetch_experiments_table(experiments=name_matches_regex & exclude_archived) ``` ### Example 5: Fetch runs belonging to specific experiment Each run's experiment information is stored in the `sys/experiment` namespace. To query runs belonging to a specific experiment, use the runs API and construct a filter on the `sys/experiment/name` attribute: ```py import neptune_query.runs as nq_runs from neptune_query.filters import Filter experiment_name_filter = Filter.eq("sys/experiment/name", "kittiwake-week-1") nq_runs.list_runs(runs=experiment_name_filter) ``` --- ## License This project is licensed under the Apache License Version 2.0. For details, see [Apache License Version 2.0][license]. [license]: http://www.apache.org/licenses/LICENSE-2.0
text/markdown
neptune.ai
contact@neptune.ai
null
null
Apache-2.0
MLOps, ML Experiment Tracking, ML Model Registry, ML Model Store, ML Metadata Store
[ "Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Developers", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Opera...
[]
null
null
<4.0,>=3.10
[]
[]
[]
[ "PyJWT<3.0.0,>=2.0.0", "attrs>=21.3.0", "azure-storage-blob<13.0.0,>=12.7.0", "httpx[http2]<0.28.2,>=0.15.4", "pandas>=1.4.0", "protobuf<7,>=4.21.1", "python-dateutil<3.0.0,>=2.8.0", "tqdm>=4.66.0" ]
[]
[]
[]
[ "Documentation, https://docs.neptune.ai/", "Homepage, https://neptune.ai/", "Repository, https://github.com/neptune-ai/neptune-query", "Tracker, https://github.com/neptune-ai/neptune-query/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:35:50.072887
neptune_query-1.12.0b1.tar.gz
126,051
fd/ea/a0bd34af638823049c5ad42e15eda2bfee0e86d5e1f5acd73ec9527a5eea/neptune_query-1.12.0b1.tar.gz
source
sdist
null
false
8d6c24694e289ef806689f122b020b3e
d74c9ae26dfc37137bac946d25af2c297c697741079c1017eee9ef76985e57b6
fdeaa0bd34af638823049c5ad42e15eda2bfee0e86d5e1f5acd73ec9527a5eea
null
[ "LICENSE" ]
227
2.4
kubernify
1.0.0
Verify Kubernetes deployments match a version manifest with deep stability auditing. Checks convergence, revision consistency, and pod health.
# kubernify [![PyPI version](https://img.shields.io/pypi/v/kubernify.svg)](https://pypi.org/project/kubernify/) [![Python versions](https://img.shields.io/pypi/pyversions/kubernify.svg)](https://pypi.org/project/kubernify/) [![CI](https://github.com/gs202/Kubernify/actions/workflows/ci.yml/badge.svg)](https://github.com/gs202/Kubernify/actions/workflows/ci.yml) [![License](https://img.shields.io/badge/license-Apache%202.0-blue.svg)](LICENSE) Verify Kubernetes deployments match a version manifest with deep stability auditing. Checks convergence, revision consistency, and pod health. --- ## Features - **Manifest-driven verification** — Provide a JSON manifest of expected versions; kubernify verifies the cluster matches - **Deep stability auditing** — Goes beyond version checks: convergence, revision consistency, pod health, DaemonSet scheduling, Job completion - **Retry-until-converged loop** — Waits for rollouts to complete rather than just snapshot-checking - **Repository-relative image parsing** — Flexible component name extraction from any image registry format - **Comprehensive workload support** — Deployments, StatefulSets, DaemonSets, Jobs, and CronJobs - **Zero-replica awareness** — Verifies version from PodSpec even when HPA/KEDA has scaled to zero - **Structured JSON reports** — Machine-readable output for CI/CD pipeline integration --- ## Installation ```bash pip install kubernify ``` Or with [pipx](https://pipx.pypa.io/) for isolated CLI usage: ```bash pipx install kubernify ``` Or with [uv](https://docs.astral.sh/uv/): ```bash uv add kubernify ``` --- ## Quick Start ```bash # Verify backend and frontend match expected versions in the "production" namespace kubernify \ --context my-cluster-context \ --anchor my-app \ --namespace production \ --manifest '{"backend": "v1.2.3", "frontend": "v1.2.4"}' ``` kubernify will connect to the cluster, discover all matching workloads, verify their image versions against the manifest, run stability audits, and exit with code `0` (pass), `1` (fail), or `2` (timeout). --- ## CLI Reference ``` kubernify [OPTIONS] ``` | Argument | Description | Default | |----------|-------------|---------| | `--context` | Kubeconfig context name. Mutually exclusive with `--gke-project`. | From kubeconfig | | `--gke-project` | GCP project ID for GKE context resolution. Mutually exclusive with `--context`. | — | | `--anchor` | **(required)** Image path anchor for component name extraction. See [How Image Anchor Works](#how-image-anchor-works). | — | | `--manifest` | **(required)** JSON version manifest, e.g. `'{"backend": "v1.2.3"}'`. | — | | `--namespace` | Kubernetes namespace to verify. | From kubeconfig context | | `--required-workloads` | Comma-separated workload name patterns that must exist. | — | | `--skip-containers` | Comma-separated container name patterns to skip during verification. | — | | `--min-uptime` | Minimum pod uptime in seconds for stability checks. | `0` | | `--restart-threshold` | Maximum allowed container restart count. | `3` | | `--timeout` | Global timeout in seconds for the verification loop. | `300` | | `--allow-zero-replicas` | Allow workloads with zero replicas to pass verification. | `false` | | `--dry-run` | Snapshot check without waiting for convergence. | `false` | | `--include-statefulsets` | Include StatefulSets in workload discovery. | `false` | | `--include-daemonsets` | Include DaemonSets in workload discovery. | `false` | | `--include-jobs` | Include Jobs and CronJobs in workload discovery. | `false` | --- ## Usage Examples ### Basic Usage — Direct Kubeconfig Context ```bash kubernify \ --context my-cluster-context \ --anchor my-app \ --namespace production \ --manifest '{"backend": "v1.2.3", "frontend": "v1.2.4"}' ``` ### GKE Shorthand — Resolve Context from GCP Project ```bash kubernify \ --gke-project my-gke-project-123456 \ --anchor my-app \ --namespace production \ --manifest '{"backend": "v1.2.3", "frontend": "v1.2.4"}' ``` ### In-Cluster — Running Inside a Kubernetes Pod ```bash # No --context needed; auto-detects in-cluster config and namespace kubernify \ --anchor my-app \ --manifest '{"backend": "v1.2.3", "frontend": "v1.2.4"}' ``` ### Full-Featured — All Options ```bash kubernify \ --context my-cluster-context \ --anchor my-app \ --namespace production \ --manifest '{"backend": "v1.2.3", "frontend": "v1.2.4", "worker": "v1.2.3"}' \ --required-workloads "backend, frontend, worker" \ --skip-containers "istio-proxy, envoy, fluent-bit" \ --include-statefulsets \ --include-daemonsets \ --include-jobs \ --min-uptime 120 \ --restart-threshold 5 \ --timeout 600 \ --allow-zero-replicas ``` ### Dry Run — Snapshot Check Without Waiting ```bash kubernify \ --context my-cluster-context \ --anchor my-app \ --manifest '{"backend": "v1.2.3"}' \ --dry-run ``` ### CI/CD Integration — GitHub Actions ```yaml jobs: verify-deployment: runs-on: ubuntu-latest steps: - name: Set up kubeconfig run: | echo "${{ secrets.KUBECONFIG }}" > /tmp/kubeconfig export KUBECONFIG=/tmp/kubeconfig - name: Install kubernify run: pip install kubernify - name: Verify deployment run: | kubernify \ --context ${{ secrets.KUBE_CONTEXT }} \ --anchor my-app \ --manifest '${{ steps.build.outputs.manifest }}' \ --timeout 600 \ --min-uptime 60 ``` --- ## Programmatic Usage kubernify can be used as a Python library for custom verification workflows: ```python from kubernify import __version__, VerificationStatus from kubernify.kubernetes_controller import KubernetesController from kubernify.workload_discovery import WorkloadDiscovery from kubernify.cli import construct_component_map, verify_versions controller = KubernetesController(context="my-cluster") discovery = WorkloadDiscovery(k8s_controller=controller) workloads, _ = discovery.discover_cluster_state(namespace="production") component_map = construct_component_map( workloads=workloads, manifest={"backend": "v1.2.3"}, repository_anchor="my-app", ) results = verify_versions(manifest={"backend": "v1.2.3"}, component_map=component_map) if results.errors: print(f"Verification failed: {results.errors}") ``` --- ## How Image Anchor Works kubernify uses a **repository-relative anchor** to extract component names from container image paths. The `--anchor` argument specifies the path segment after which the component name is derived. ``` Image: registry.example.com/my-org/my-app/backend:v1.2.3 └─ registry ──────────┘└─ org ┘└anchor┘└component┘ ↓ --anchor my-app → component = "backend" ``` **More examples:** | Image | `--anchor` | Extracted Component | |-------|-----------|-------------------| | `registry.example.com/my-org/my-app/backend:v1.2.3` | `my-app` | `backend` | | `registry.example.com/my-org/my-app/api/server:v2.0.0` | `my-app` | `api/server` | | `gcr.io/my-project/my-app/worker:v1.0.0` | `my-app` | `worker` | The extracted component name is then matched against the keys in your `--manifest` JSON to verify the correct version is deployed. --- ## Architecture ```mermaid graph TD A[CLI Entry Point] --> B[Argument Parser] B --> C{Context Mode} C -->|--context| D[Direct kubeconfig context] C -->|--gke-project| E[GKE context resolver] C -->|neither| F[In-cluster or default kubeconfig] D --> G[KubernetesController] E --> G F --> G G --> H[WorkloadDiscovery] H --> I[Fetch Deployments/StatefulSets/DaemonSets/Jobs] H --> J[Inspect Workloads - concurrent] J --> K[Image Parser] K --> L[Component Map Construction] L --> M[Version Verification] J --> N[StabilityAuditor] N --> O[Convergence Check] N --> P[Revision Consistency] N --> Q[Pod Health] N --> R[DaemonSet Scheduling] N --> S[Job Completion] M --> T[Report Generator] N --> T T --> U[JSON Report Output] U --> V{Exit Code} V -->|0| W[PASS] V -->|1| X[FAIL] V -->|2| Y[TIMEOUT] ``` --- ## Exit Codes | Code | Meaning | Description | |------|---------|-------------| | `0` | **PASS** | All workloads match the manifest and pass stability audits | | `1` | **FAIL** | One or more workloads have version mismatches or stability issues | | `2` | **TIMEOUT** | Verification did not converge within the `--timeout` window | --- ## Prerequisites ### Python - Python **>= 3.10** ### For GKE Users If using `--gke-project` for automatic GKE context resolution: 1. Install the [Google Cloud SDK](https://cloud.google.com/sdk/docs/install) 2. Install the GKE auth plugin: ```bash gcloud components install gke-gcloud-auth-plugin ``` 3. Authenticate: ```bash gcloud auth login gcloud container clusters get-credentials CLUSTER_NAME --project PROJECT_ID ``` ### RBAC Permissions kubernify requires **read-only** access to workloads and pods. Apply the following RBAC configuration: ```yaml apiVersion: rbac.authorization.k8s.io/v1 kind: Role metadata: name: kubernify-reader namespace: <namespace> rules: - apiGroups: ["apps"] resources: ["deployments", "statefulsets", "daemonsets", "replicasets"] verbs: ["get", "list"] - apiGroups: ["batch"] resources: ["jobs", "cronjobs"] verbs: ["get", "list"] - apiGroups: [""] resources: ["pods"] verbs: ["get", "list"] --- apiVersion: rbac.authorization.k8s.io/v1 kind: RoleBinding metadata: name: kubernify-reader-binding namespace: <namespace> subjects: - kind: ServiceAccount name: kubernify namespace: <namespace> roleRef: kind: Role name: kubernify-reader apiGroup: rbac.authorization.k8s.io ``` --- ## Contributing Contributions are welcome! Please see [CONTRIBUTING.md](CONTRIBUTING.md) for development setup, coding standards, and the PR process. --- ## License This project is licensed under the Apache License 2.0 — see the [LICENSE](LICENSE) file for details.
text/markdown
null
gs202 <gs202@users.noreply.github.com>
null
null
Apache-2.0
deployment, devops, k8s, kubernetes, verification, version
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Software ...
[]
null
null
>=3.10
[]
[]
[]
[ "kubernetes>=28.1.0" ]
[]
[]
[]
[ "Homepage, https://github.com/gs202/Kubernify", "Documentation, https://github.com/gs202/Kubernify#readme", "Issues, https://github.com/gs202/Kubernify/issues" ]
uv/0.10.4 {"installer":{"name":"uv","version":"0.10.4","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
2026-02-19T12:34:15.026658
kubernify-1.0.0.tar.gz
101,439
e3/41/8223a33b857e2ee0ea2b261538b48439b75cbba805ec89e4e492445c7ab9/kubernify-1.0.0.tar.gz
source
sdist
null
false
de3950836e049930aae20e2957b9933a
0d08da73a08afa3c4811867f8a9c65b55e17ba99814c634809af2f31acc44986
e3418223a33b857e2ee0ea2b261538b48439b75cbba805ec89e4e492445c7ab9
null
[ "LICENSE" ]
288
2.4
influxdb3-python
0.18.0
Community Python client for InfluxDB 3.0
<!--home-start--> <p align="center"> <img src="https://github.com/InfluxCommunity/influxdb3-python/blob/main/python-logo.png?raw=true" alt="Your Image" width="150px"> </p> <p align="center"> <a href="https://influxdb3-python.readthedocs.io/en/latest/"> <img src="https://img.shields.io/readthedocs/influxdb3-python/latest" alt="Readthedocs document"> </a> <a href="https://pypi.org/project/influxdb3-python/"> <img src="https://img.shields.io/pypi/v/influxdb3-python.svg" alt="PyPI version"> </a> <a href="https://pypi.org/project/influxdb3-python/"> <img src="https://img.shields.io/pypi/dm/influxdb3-python.svg" alt="PyPI downloads"> </a> <a href="https://github.com/InfluxCommunity/influxdb3-python/actions/workflows/codeql-analysis.yml"> <img src="https://github.com/InfluxCommunity/influxdb3-python/actions/workflows/codeql-analysis.yml/badge.svg?branch=main" alt="CodeQL analysis"> </a> <a href="https://dl.circleci.com/status-badge/redirect/gh/InfluxCommunity/influxdb3-python/tree/main"> <img src="https://dl.circleci.com/status-badge/img/gh/InfluxCommunity/influxdb3-python/tree/main.svg?style=svg" alt="CircleCI"> </a> <a href="https://codecov.io/gh/InfluxCommunity/influxdb3-python"> <img src="https://codecov.io/gh/InfluxCommunity/influxdb3-python/branch/main/graph/badge.svg" alt="Code Cov"/> </a> <a href="https://influxcommunity.slack.com"> <img src="https://img.shields.io/badge/slack-join_chat-white.svg?logo=slack&style=social" alt="Community Slack"> </a> </p> # InfluxDB 3.0 Python Client ## Introduction `influxdb_client_3` is a Python module that provides a simple and convenient way to interact with InfluxDB 3.0. This module supports both writing data to InfluxDB and querying data using the Flight client, which allows you to execute SQL and InfluxQL queries on InfluxDB 3.0. We offer a ["Getting Started: InfluxDB 3.0 Python Client Library"](https://www.youtube.com/watch?v=tpdONTm1GC8) video that goes over how to use the library and goes over the examples. ## Dependencies - `pyarrow` (automatically installed) - `pandas` (optional) ## Installation You can install 'influxdb3-python' using `pip`: ```bash pip install influxdb3-python ``` Note: This does not include Pandas support. If you would like to use key features such as `to_pandas()` and `write_file()` you will need to install `pandas` separately. *Note: Please make sure you are using 3.6 or above. For the best performance use 3.11+* # Usage One of the easiest ways to get started is to checkout the ["Pokemon Trainer Cookbook"](https://github.com/InfluxCommunity/influxdb3-python/blob/main/Examples/pokemon-trainer/cookbook.ipynb). This scenario takes you through the basics of both the client library and Pyarrow. ## Importing the Module ```python from influxdb_client_3 import InfluxDBClient3, Point ``` ## Initialization If you are using InfluxDB Cloud, then you should note that: 1. Use bucket name for `database` or `bucket` in function argument. ```python client = InfluxDBClient3(token="your-token", host="your-host", database="your-database") ``` ## Writing Data You can write data using the Point class, or supplying line protocol. ### Using Points ```python point = Point("measurement").tag("location", "london").field("temperature", 42) client.write(point) ``` ### Using Line Protocol ```python point = "measurement fieldname=0" client.write(point) ``` ### Write from file Users can import data from CSV, JSON, Feather, ORC, Parquet ```python import influxdb_client_3 as InfluxDBClient3 import pandas as pd import numpy as np from influxdb_client_3 import write_client_options, WritePrecision, WriteOptions, InfluxDBError class BatchingCallback(object): def __init__(self): self.write_count = 0 def success(self, conf, data: str): self.write_count += 1 print(f"Written batch: {conf}, data: {data}") def error(self, conf, data: str, exception: InfluxDBError): print(f"Cannot write batch: {conf}, data: {data} due: {exception}") def retry(self, conf, data: str, exception: InfluxDBError): print(f"Retryable error occurs for batch: {conf}, data: {data} retry: {exception}") callback = BatchingCallback() write_options = WriteOptions(batch_size=100, flush_interval=10_000, jitter_interval=2_000, retry_interval=5_000, max_retries=5, max_retry_delay=30_000, exponential_base=2) wco = write_client_options(success_callback=callback.success, error_callback=callback.error, retry_callback=callback.retry, write_options=write_options ) with InfluxDBClient3.InfluxDBClient3( token="INSERT_TOKEN", host="eu-central-1-1.aws.cloud2.influxdata.com", database="python", write_client_options=wco) as client: client.write_file( file='./out.csv', timestamp_column='time', tag_columns=["provider", "machineID"]) print(f'DONE writing from csv in {callback.write_count} batch(es)') ``` ### Pandas DataFrame ```python import pandas as pd # Create a DataFrame with a timestamp column df = pd.DataFrame({ 'time': pd.to_datetime(['2024-01-01', '2024-01-02', '2024-01-03']), 'trainer': ['Ash', 'Misty', 'Brock'], 'pokemon_id': [25, 120, 74], 'pokemon_name': ['Pikachu', 'Staryu', 'Geodude'] }) # Write the DataFrame - timestamp_column is required for consistency client.write_dataframe( df, measurement='caught', timestamp_column='time', tags=['trainer', 'pokemon_id'] ) ``` ### Polars DataFrame ```python import polars as pl # Create a DataFrame with a timestamp column df = pl.DataFrame({ 'time': ['2024-01-01T00:00:00Z', '2024-01-02T00:00:00Z'], 'trainer': ['Ash', 'Misty'], 'pokemon_id': [25, 120], 'pokemon_name': ['Pikachu', 'Staryu'] }) # Write the DataFrame - same API works for both pandas and polars client.write_dataframe( df, measurement='caught', timestamp_column='time', tags=['trainer', 'pokemon_id'] ) ``` ## Querying ### Querying with SQL ```python query = "select * from measurement" reader = client.query(query=query, language="sql") table = reader.read_all() print(table.to_pandas().to_markdown()) ``` ### Querying to DataFrame ```python # Query directly to a pandas DataFrame (default) df = client.query_dataframe("SELECT * FROM caught WHERE trainer = 'Ash'") # Query to a polars DataFrame df = client.query_dataframe("SELECT * FROM caught", frame_type="polars") ``` ### Querying with influxql ```python query = "select * from measurement" reader = client.query(query=query, language="influxql") table = reader.read_all() print(table.to_pandas().to_markdown()) ``` ### gRPC compression #### Request compression Request compression is not supported by InfluxDB 3 — the client sends uncompressed requests. #### Response compression Response compression is enabled by default. The client sends the `grpc-accept-encoding: identity, deflate, gzip` header, and the server returns gzip-compressed responses (if supported). The client automatically decompresses them — no configuration required. To **disable response compression**: ```python # Via constructor parameter client = InfluxDBClient3( host="your-host", token="your-token", database="your-database", disable_grpc_compression=True ) # Or via environment variable # INFLUX_DISABLE_GRPC_COMPRESSION=true client = InfluxDBClient3.from_env() ``` ## Windows Users Currently, Windows users require an extra installation when querying via Flight natively. This is due to the fact gRPC cannot locate Windows root certificates. To work around this please follow these steps: Install `certifi` ```bash pip install certifi ``` Next include certifi within the flight client options: ```python import certifi import influxdb_client_3 as InfluxDBClient3 from influxdb_client_3 import flight_client_options fh = open(certifi.where(), "r") cert = fh.read() fh.close() client = InfluxDBClient3.InfluxDBClient3( token="", host="b0c7cce5-8dbc-428e-98c6-7f996fb96467.a.influxdb.io", database="flightdemo", flight_client_options=flight_client_options( tls_root_certs=cert)) table = client.query( query="SELECT * FROM flight WHERE time > now() - 4h", language="influxql") print(table.to_pandas()) ``` You may include your own root certificate in this manner as well. If connecting to InfluxDB fails with error `DNS resolution failed` when using domain name, example `www.mydomain.com`, then try to set environment variable `GRPC_DNS_RESOLVER=native` to see if it works. # Contributing Tests are run using `pytest`. ```bash # Clone the repository git clone https://github.com/InfluxCommunity/influxdb3-python cd influxdb3-python # Create a virtual environment and activate it python3 -m venv .venv source .venv/bin/activate # Install the package and its dependencies pip install -e .[pandas,polars,dataframe,test] # Run the tests python -m pytest . ``` <!--home-end-->
text/markdown
InfluxData
contact@influxdata.com
null
null
null
null
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: ...
[]
https://github.com/InfluxCommunity/influxdb3-python
null
>=3.8
[]
[]
[]
[ "reactivex>=4.0.4", "certifi>=14.05.14", "python_dateutil>=2.5.3", "urllib3>=1.26.0", "pyarrow>=8.0.0", "pandas; extra == \"pandas\"", "polars; extra == \"polars\"", "pandas; extra == \"dataframe\"", "polars; extra == \"dataframe\"", "pytest; extra == \"test\"", "pytest-cov; extra == \"test\"", ...
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:32:22.017340
influxdb3_python-0.18.0.tar.gz
98,718
4d/d9/bf0a0012ac01aa8addb49df4860f2f4aa592ac37f42e2d1e8ecaea6412d5/influxdb3_python-0.18.0.tar.gz
source
sdist
null
false
beb720a511ac50596fea266ee9ba36d5
9ffe285fc9711820264f77ba4a03524df5cddbab3ebfa90967e88db2e89fbba6
4dd9bf0a0012ac01aa8addb49df4860f2f4aa592ac37f42e2d1e8ecaea6412d5
null
[ "LICENSE" ]
9,665
2.4
PyPFD
2026.0.0.2
PFDavg calculation
Introduction PyPFD is a Python library designed to calculate the Probability of Failure on Demand (PFD) in accordance with the international safety standards IEC 61508 and IEC 61511. It provides a way to estimate the reliability of Safety Devices, making it easier for engineers and safety professionals to perform consistent SIS assessments. It allows you to evaluate PFDavg for various architectures (1oo1, 1oo2, 2oo2, 2oo3, 1oo3, and KooN using a general formula). The library provides the following formulas: PyPFDRBDAvg: pfd_RBD_avg_1oo1(λ_du, λ_dd, T1_month, MTTR) pfd_RBD_avg_1oo1_pt(λ_du, λ_dd, T1_month, T2_month, PDC, MTTR) pfd_RBD_avg_1oo2(λ_du, λ_dd, β, βd, T1_month, MTTR) pfd_RBD_avg_1oo2_pt(λ_du, λ_dd, β, βd, T1_month, T2_month, PDC, MTTR) pfd_RBD_avg_1oo3(λ_du, λ_dd, β, βd, T1_month, MTTR) pfd_RBD_avg_2oo2(λ_du, λ_dd, T1_month, MTTR) pfd_RBD_avg_2oo2_pt(λ_du, λ_dd, T1_month, T2_month, PDC, MTTR) pfd_RBD_avg_2oo3(λ_du, λ_dd, β, βd, T1_month, MTTR) pfd_RBD_avg_KooN(K, N, λ_du, λ_dd, β, βd, T1_month, MTTR) PyPFDMarkov: pfd_Mkv_avg_1oo1_2pt(λ_du: float,λ_dd: float,λ_s: float,T_pt1_month: float,T_pt2_month: float,T1_month: float,PDC1: float,PDC2: float,MTTR: float) pfd_Mkv_avg_1oo2(λ_du: float,λ_dd: float,λ_su: float,λ_sd: float,β:float, βd:float,T1_month: float,MTTR:float) pfd_plot_Mkv_avg_1oo1_2pt(λ_du: float,λ_dd: float,λ_s: float,T_pt1_month: float,T_pt2_month: float,T1_month: float,PDC1: float,PDC2: float,MTTR: float,interval:int) pfd_plot_Mkv_avg_1oo2(λ_du: float,λ_dd: float,λ_su: float,λ_sd: float,β:float, βd:float,T1_month: float,MTTR:float,interval:int) PyPFDN2595: pfd_N2595_avg_MooN(M:int,N:int,λ_du:float,λ_dd:float,β:float,T1_month:float,MTTR:float) -> float: pfd_N2595_avg_1oo1_1pt(λ_du:float,λ_dd:float,PTC:float,TI_month:float,MT_month:float,MTTR:float) -> float: pfd_N2595_avg_1oo2_1pt(λ_du:float,λ_dd:float,β:float,TI_month:float,MT_month:float,PTC:float,MTTR:float) -> float: pfd_N2595_avg_2oo2_1pt(λ_du:float,λ_dd:float,PTC:float,TI_month:float,MT_month:float,MTTR:float) -> float: pfd_N2595_avg_2oo3_1pt(λ_du:float,λ_dd:float,β:float,TI_month:float,MT_month:float,PTC:float,MTTR:float) -> float: Parameters: λ_du = dangerous undetected failure rate per hour λ_dd = dangerous detected failure rate per hour β = common cause for safe failure in % βd = common cause for unsafe failure in % T1_month = test interval in months (with PDC effectiveness) T2_month or MT_month = test interval in months for "as good as new" condition PDC = partial diagnostic coverage of the test (% capable of revealing dangerous undetected failures) MTTR = mean time to repair All formulas assume identical devices. For combinations of different devices or different test intervals, see the formulas below (currently in development and not validated yet): def pfhKooN(K, N, λ_d, β, T1_month): def pfd_avg_1oo2_dif(λ_du1, λ_dd1, T1_month1, MTTR1, β1, βd1, λ_du2, λ_dd2, T1_month2, MTTR2, β2, βd2): def pfd_avg_2oo3_dif(λ_du1, λ_dd1, T1_month1, MTTR1, β1, βd1, λ_du2, λ_dd2, T1_month2, MTTR2, β2, βd2, λ_du3, λ_dd3, T1_month3, MTTR3, β3, βd3): def pfd_avg_1oo3_dif(λ_du1, λ_dd1, T1_month1, MTTR1, β1, βd1, λ_du2, λ_dd2, T1_month2, MTTR2, β2, βd2, λ_du3, λ_dd3, T1_month3, MTTR3, β3, βd3): Roadmap -Test and validate all existing formulas. -Create a GitHub repository explaining the logic behind the formulas. -Develop default reliability data for common devices (Analog Transmitters, Valves, and Logic Solvers). Highlights -These formulas, combined with xlwings Lite in Excel, provide an efficient and user-friendly way to perform SIS assessments. -If you need a specific architecture not present in the library, feel free to contact me for assistance.
text/markdown
null
Rafael Rocha <rafa.rocha@gmail.com>
null
null
null
SIS, SIF, IEC61508, IEC61511, PFD, reliability, functional safety, 1oo1, 1oo2, 2oo3, 2oo2, 1oo3, N2595
[ "Programming Language :: Python :: 3", "Operating System :: OS Independent" ]
[]
null
null
>=3.8
[]
[]
[]
[]
[]
[]
[]
[]
twine/6.2.0 CPython/3.13.7
2026-02-19T12:31:24.090045
pypfd-2026.0.0.2.tar.gz
8,306
7a/ea/da519a1bb2b4a6d99ec4d5ff090ce52104647035db8cec3d4780984a871e/pypfd-2026.0.0.2.tar.gz
source
sdist
null
false
38e0d77caac8f8e0b665e0c563a92a09
4c7ab2f54a6422307c5d17e8e9da30652e14743454e91f3a67353bd8765e4108
7aeada519a1bb2b4a6d99ec4d5ff090ce52104647035db8cec3d4780984a871e
null
[ "LICENSE" ]
0
2.4
copex
2.6.1
Copilot Extended - Resilient wrapper for GitHub Copilot SDK with auto-retry, Ralph Wiggum loops, and more
# Copex — Copilot Extended [![PyPI](https://img.shields.io/pypi/v/copex)](https://pypi.org/project/copex/) [![Python 3.10+](https://img.shields.io/badge/python-3.10%2B-blue)](https://python.org) [![License: MIT](https://img.shields.io/badge/license-MIT-green)](LICENSE) A resilient Python wrapper for the GitHub Copilot SDK with automatic retry, Ralph Wiggum loops, fleet orchestration, and a beautiful CLI. ## Features - **Automatic retry** with adaptive exponential backoff and jitter - **Circuit breaker** (sliding-window) to avoid hammering a failing backend - **Model fallback chains** — automatically try the next model when one fails - **Session pooling** with LRU eviction and pre-warming - **Ralph Wiggum loops** — iterative AI development via repeated prompts - **Fleet mode** — run multiple tasks in parallel with dependency ordering - **Council mode** — multi-model deliberation with a chair model - **Plan mode** — AI-generated step-by-step plans with execution & checkpointing - **MCP integration** — connect external Model Context Protocol servers - **Skill discovery** — auto-discover and load skill files from repo/user dirs - **Beautiful CLI** with Rich terminal output, themes, and streaming - **CLI client mode** — bypass the SDK to access all models via subprocess - **Metrics & cost tracking** — token usage, timing, success rates ## Installation ```bash pip install copex ``` Or with [pipx](https://pipx.pypa.io/) for isolated CLI usage: ```bash pipx install copex ``` **Prerequisite:** The GitHub Copilot CLI must be installed and authenticated: ```bash npm i -g @github/copilot copilot login ``` ## Quick Start ### CLI ```bash # One-shot prompt copex chat "Explain the builder pattern in Python" # Interactive session copex interactive # Pipe from stdin echo "What is a monad?" | copex chat --stdin # Choose model and reasoning copex chat "Optimize this SQL" -m gpt-5.2-codex -r xhigh ``` ### Python API ```python import asyncio from copex import Copex, CopexConfig, Model, ReasoningEffort async def main(): config = CopexConfig( model=Model.CLAUDE_OPUS_4_6, reasoning_effort=ReasoningEffort.HIGH, ) async with Copex(config) as client: response = await client.send("Explain quicksort") print(response.content) asyncio.run(main()) ``` #### Streaming ```python async with Copex(config) as client: async for chunk in client.stream("Write a prime sieve"): if chunk.type == "message": print(chunk.delta, end="", flush=True) elif chunk.type == "reasoning": print(f"\033[2m{chunk.delta}\033[0m", end="", flush=True) ``` ## Configuration Copex loads configuration from `~/.config/copex/config.toml` automatically. Generate a starter config with: ```bash copex init ``` ### Config file example ```toml model = "claude-opus-4.6" reasoning_effort = "high" streaming = true use_cli = false timeout = 300.0 auto_continue = true ui_theme = "default" # default, midnight, mono, sunset ui_density = "extended" # compact, extended [retry] max_retries = 5 base_delay = 1.0 max_delay = 30.0 exponential_base = 2.0 # Skills skills = ["code-review"] # skill_directories = ["/path/to/skills"] # disabled_skills = ["some-skill"] # MCP servers # mcp_config_file = "~/.config/copex/mcp.json" # Tool filtering # available_tools = ["bash", "view"] # excluded_tools = ["dangerous-tool"] ``` ### Environment variables | Variable | Description | |---|---| | `COPEX_MODEL` | Override the default model | | `COPEX_REASONING` | Override the reasoning effort | ## CLI Commands | Command | Description | |---|---| | `copex chat` | Send a prompt with automatic retry | | `copex interactive` | Start an interactive chat session | | `copex tui` | Launch the full TUI interface | | `copex ralph` | Start a Ralph Wiggum iterative loop | | `copex plan` | Generate and execute step-by-step plans | | `copex fleet` | Run multiple tasks in parallel | | `copex council` | Multi-model deliberation on a task | | `copex models` | List available models | | `copex skills list` | List discovered skills | | `copex skills show` | Show skill content | | `copex render` | Render a JSONL session log | | `copex status` | Show auth and CLI status | | `copex config` | Show/edit configuration | | `copex init` | Generate a starter config file | | `copex login` | Authenticate with GitHub Copilot | | `copex logout` | Remove authentication | | `copex completions` | Generate shell completion scripts | ### Common flags ``` -m, --model Model to use -r, --reasoning Reasoning effort (none, low, medium, high, xhigh) -c, --config Config file path -S, --skill-dir Add skill directory --use-cli Use CLI subprocess instead of SDK --json Machine-readable JSON output -q, --quiet Content only, no formatting ``` ## Models Copex supports all models available through the Copilot SDK: | Model | Reasoning | xhigh | |---|---|---| | `gpt-5.2-codex` | ✅ | ✅ | | `gpt-5.2` | ✅ | ✅ | | `gpt-5.1-codex` | ✅ | ❌ | | `gpt-5.1-codex-max` | ✅ | ❌ | | `gpt-5.1-codex-mini` | ✅ | ❌ | | `gpt-5.1` | ✅ | ❌ | | `gpt-5` | ✅ | ❌ | | `gpt-5-mini` | ✅ | ❌ | | `gpt-4.1` | ❌ | ❌ | | `claude-opus-4.6` | ✅ | ❌ | | `claude-opus-4.6-fast` | ✅ | ❌ | | `claude-opus-4.5` | ❌ | ❌ | | `claude-sonnet-4.5` | ❌ | ❌ | | `claude-sonnet-4` | ❌ | ❌ | | `claude-haiku-4.5` | ❌ | ❌ | | `gemini-3-pro-preview` | ❌ | ❌ | Copex also discovers models dynamically from `copilot --help` at runtime, so newly added models work automatically. ```bash copex models # List all available models ``` ## Reasoning Effort Five levels control how much thinking the model does: | Level | Description | |---|---| | `none` | No extended reasoning | | `low` | Minimal reasoning | | `medium` | Balanced | | `high` | Thorough reasoning (default) | | `xhigh` | Maximum reasoning (GPT-5.2+ only) | If you request an unsupported level, Copex automatically downgrades and warns you. ## Advanced Features ### Retry & Backoff Copex uses adaptive per-error-category backoff strategies: ```python config = CopexConfig( retry=RetryConfig( max_retries=5, base_delay=1.0, max_delay=30.0, exponential_base=2.0, ), ) ``` Error categories (rate limit, network, server, auth, client) each have their own backoff curve. Rate limit errors respect `Retry-After` headers when available. ### Circuit Breaker A sliding-window circuit breaker opens when the failure rate exceeds 50% in the last 10 requests, then enters a 60-second cooldown before retrying. ### Model Fallback Chains ```python client = Copex(config, fallback_chain=["claude-opus-4.6", "gpt-5.2-codex", "gpt-5.1-codex"]) ``` If the primary model fails, Copex transparently tries the next model in the chain. ### Session Pooling ```python from copex.client import SessionPool pool = SessionPool(max_sessions=5, max_idle_time=300) async with pool.acquire(client, config) as session: await session.send({"prompt": "Hello"}) ``` ### Ralph Wiggum Loops Iterative AI development: the same prompt is fed repeatedly, with the AI seeing its own previous work to iteratively improve. ```bash copex ralph "Build a REST API with CRUD and tests" \ --promise "ALL TESTS PASSING" \ -n 20 ``` ```python from copex.ralph import RalphWiggum ralph = RalphWiggum(copex_client) result = await ralph.loop( prompt="Build a REST API with CRUD operations", completion_promise="API COMPLETE", max_iterations=30, ) ``` ### Plan Mode AI-generated step-by-step execution with checkpointing and resume: ```bash copex plan "Build a REST API" --execute copex plan "Build a REST API" --review # Confirm before executing copex plan --resume # Resume from checkpoint copex plan "Build a REST API" --visualize ascii ``` ### Fleet Mode Run multiple tasks in parallel with optional dependency ordering: ```bash copex fleet "Write tests" "Fix linting" "Update docs" --max-concurrent 3 copex fleet --file tasks.toml ``` ```python from copex import Fleet, FleetConfig async with Fleet(config) as fleet: fleet.add("Write auth tests") fleet.add("Refactor DB", depends_on=["write-auth-tests"]) results = await fleet.run() ``` Features: adaptive concurrency, rate-limit backoff, git finalize, artifact export. ### Council Mode Multi-model deliberation — multiple models investigate a problem, then a chair model synthesizes the best solution: ```bash copex council "Design a caching strategy for our API" \ --chair-model claude-opus-4.6 \ --codex-model gpt-5.2-codex \ --gemini-model gemini-3-pro-preview ``` ### MCP Integration Connect external Model Context Protocol servers: ```toml # In config.toml mcp_config_file = "~/.config/copex/mcp.json" ``` Or pass inline: ```bash copex fleet --mcp-config mcp.json "Analyze the codebase" ``` ### Checkpointing & Persistence Ralph loops and plan execution save checkpoints to disk for crash recovery. Sessions can be saved and restored across runs. ## CLI Client Mode Use `--use-cli` to bypass the SDK and invoke the Copilot CLI directly as a subprocess. This is useful when the SDK doesn't support a model but the CLI does: ```bash copex chat "Hello" --use-cli -m claude-opus-4.6 ``` ```python config = CopexConfig(use_cli=True, model=Model.CLAUDE_OPUS_4_6) client = make_client(config) # Returns CopilotCLI instead of Copex ``` ## Development ```bash # Clone and install git clone https://github.com/Arthur742Ramos/copex.git cd copex pip install -e ".[dev]" # Run tests python -m pytest tests/ -v # Run with coverage python -m pytest tests/ --cov=copex --cov-report=term-missing # Lint ruff check src/ ruff format src/ ``` ## License [MIT](LICENSE)
text/markdown
null
Arthur Ramos <arthur742ramos@users.noreply.github.com>
null
null
null
ai, copex, copilot, github, ralph-wiggum, retry, sdk
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12" ]
[]
null
null
>=3.10
[]
[]
[]
[ "github-copilot-sdk>=0.1.21", "prompt-toolkit>=3.0.0", "pydantic>=2.0.0", "rich>=13.0.0", "tomli-w>=1.0.0", "tomli; python_version < \"3.11\"", "typer>=0.9.0", "pytest; extra == \"dev\"", "pytest-asyncio; extra == \"dev\"", "pytest-cov; extra == \"dev\"", "ruff; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://github.com/Arthur742Ramos/copex", "Repository, https://github.com/Arthur742Ramos/copex", "Issues, https://github.com/Arthur742Ramos/copex/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:31:02.355161
copex-2.6.1.tar.gz
185,460
13/74/cf5e4fc9121f24556e189ae308603d3268b8464169816b2debf21de38fa6/copex-2.6.1.tar.gz
source
sdist
null
false
d327e8fd0613bef7cbf83c8a1e2b78b3
aac54ebfc1706ceda268f86cbd7c512f1c645a0241748f009e1ee20916f250ee
1374cf5e4fc9121f24556e189ae308603d3268b8464169816b2debf21de38fa6
MIT
[ "LICENSE" ]
275
2.4
onecode-pycg
1.2.0
PyCG - Practical Python Call Graphs
# :warning: Notes Forked from https://github.com/vitsalis/pycg Changes: See https://github.com/deeplime-io/PyCG/tree/onecode Essentially added code tracks the order of the calls and the code associated to it. It is used by OneCode to properly interpret code based on the **excellent** `PyCG`. Why a new PyPi package? Well, the not-so-great PyPi doesn't allow to have forked public repositories as part of the dependencies. Nice right? # PyCG - Practical Python Call Graphs [![Tests](https://github.com/deeplime-io/PyCG/actions/workflows/test.yaml/badge.svg)](https://github.com/vitsalis/PyCG/actions/workflows/test.yaml) PyCG generates call graphs for Python code using static analysis. It efficiently supports * Higher order functions * Twisted class inheritance schemes * Automatic discovery of imported modules for further analysis * Nested definitions You can read the full methodology as well as a complete evaluation on the [ICSE 2021 paper](https://arxiv.org/pdf/2103.00587.pdf). You can cite PyCG as follows. Vitalis Salis, Thodoris Sotiropoulos, Panos Louridas, Diomidis Spinellis and Dimitris Mitropoulos. PyCG: Practical Call Graph Generation in Python. In _43rd International Conference on Software Engineering, ICSE '21_, 25–28 May 2021. > **PyCG** is archived. Due to limited availability, no further development > improvements are planned. Happy to help anyone that wants to create a fork to > continue development. # Installation PyCG is implemented in Python3 and requires Python version 3.4 or higher. It also has no dependencies. Simply: ``` pip install onecode-pycg ``` # Usage ``` ~ >>> pycg -h usage: __main__.py [-h] [--package PACKAGE] [--product PRODUCT] [--forge FORGE] [--version VERSION] [--timestamp TIMESTAMP] [--max-iter MAX_ITER] [--operation {call-graph,key-error}] [--as-graph-output AS_GRAPH_OUTPUT] [-o OUTPUT] [entry_point ...] positional arguments: entry_point Entry points to be processed optional arguments: -h, --help show this help message and exit --package PACKAGE Package containing the code to be analyzed --max-iter MAX_ITER Maximum number of iterations through source code. If not specified a fix-point iteration will be performed. --operation {call-graph,key-error} Operation to perform. Choose call-graph for call graph generation (default) or key-error for key error detection on dictionaries. --as-graph-output AS_GRAPH_OUTPUT Output for the assignment graph -o OUTPUT, --output OUTPUT Output path ``` # Call Graph Output ## Simple JSON format The call edges are in the form of an adjacency list where an edge `(src, dst)` is represented as an entry of `dst` in the list assigned to key `src`: ``` { "node1": ["node2", "node3"], "node2": ["node3"], "node3": [] } ``` ## FASTEN Format Dropped - not useful for OneCode and requires porting of `pkg_resources` For an up-to-date description of the FASTEN format refer to the [FASTEN wiki](https://github.com/fasten-project/fasten/wiki/Extended-Revision-Call-Graph-format#python). # Key Errors Output We are currently experimenting on identifying potential invalid dictionary accesses on Python dictionaries (key errors). The output format for key errors is a list of dictionaries containing: - The file name in which the key error was identified - The line number inside the file - The namespace of the accessed dictionary - The key used to access the dictionary ``` [{ "filename": "mod.py", "lineno": 2, "namespace": "mod.<dict1>", "key": "key2" }, { "filename": "mod.py", "lineno": 8, "namespace": "mod.<dict1>", "key": "nokey" }] ``` # Examples All the entry points are known and we want the simple JSON format ``` ~ >>> pycg --package pkg_root pkg_root/module1.py pkg_root/subpackage/module2.py -o cg.json ``` All entry points are not known and we want the simple JSON format ``` ~ >>> pycg --package django $(find django -type f -name "*.py") -o django.json ``` # Running Tests From the root directory, first install the [mock](https://pypi.org/project/mock/) package: ``` pip3 install mock ``` Τhen, simply run the tests by executing: ``` make test ```
text/markdown
null
Vitalis Salis <vitsalis@gmail.com>
null
null
null
null
[ "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3" ]
[]
null
null
>=3.4
[]
[]
[]
[ "black>=22.12.0; extra == \"dev\"", "flake8>=6.0.0; extra == \"dev\"", "isort>=5.12.0; extra == \"dev\"", "mock; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://github.com/vitsalis/PyCG", "Bug Tracker, https://github.com/vitsalis/PyCG/issues" ]
uv/0.10.4 {"installer":{"name":"uv","version":"0.10.4","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Pop!_OS","version":"22.04","id":"jammy","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}
2026-02-19T12:30:44.328898
onecode_pycg-1.2.0.tar.gz
54,672
05/ae/8e96a51bface1acf2bd941253c390b04df5d033b7341b7d945836a3da0c0/onecode_pycg-1.2.0.tar.gz
source
sdist
null
false
0b3e1c403d89596b21dae75f2706a9de
d05841f96f0257620279df8cc46aae80575aed0658554543b77e57a954a7ba0c
05ae8e96a51bface1acf2bd941253c390b04df5d033b7341b7d945836a3da0c0
null
[ "LICENCE" ]
490
2.4
bepo
1.3.0
Detect duplicate PRs in GitHub repos
# bepo Detect duplicate pull requests in GitHub repos. No ML, no embeddings, no API keys. Just static analysis of diffs. A maintainer with 100 open PRs can run `bepo check --repo foo/bar` and in 5 minutes get a ranked list of "you should look at these pairs." That saves hours of manual review. ## The Problem Large repos waste engineering time on duplicate PRs. When multiple contributors fix the same bug independently, only one PR gets merged — the rest is wasted effort. **This actually happens.** We analyzed 100 PRs from [OpenClaw](https://github.com/openclaw/openclaw) and found: | Cluster | PRs | What happened | |---------|-----|---------------| | Matrix startup bug | **4 PRs** | 4 engineers independently fixed `startupGraceMs = 0` → `5000` | | Media token regex | 2 PRs | Identical fix submitted twice | | Feishu bitable config | 2 PRs | Same multi-account config fix | **8 duplicate PRs across 3 bug fixes.** That's real engineering time wasted. ## Proof: OpenClaw Analysis ### 30-day window (recommended) ``` $ bepo check --repo openclaw/openclaw --since 30d Found 52 potential duplicates: #20472 <-> #20491 Similarity: 100% Reason: Both fix #20468 ← Same Nextcloud Talk restart bug, two fixes #19595 <-> #19624 Similarity: 90% Reason: Both fix #19574 ← Identical PR titles, same elevatedDefault bug #20419 <-> #20441 Similarity: 83% Reason: Both fix #20410 ← Same WebChat markdown rendering fix #19865 <-> #19945 Similarity: 97% Reason: Same code: 100 lines overlap ← Two embedding provider PRs duplicating core logic #19770 <-> #20317 Similarity: 87% Reason: Same code: 9 lines overlap ← Same "hide tool calls" UI toggle, added twice ``` **Precision: ~88%** — verified by manual classification of all 52 pairs. The remaining ~12% are concurrent PRs touching the same structural code — two provider integrations sharing schema boilerplate, two locale additions hitting the same type. The kind of overlap a reviewer would want to know about. ### Full backlog (3,000 PRs) ``` $ bepo check --repo openclaw/openclaw --limit 3000 Analyzed 3000 PRs in 9.8s Found 1022 potential duplicates: #17518 <-> #17653 Similarity: 100% Reason: Both fix #17499 ← Identical browser dialog fix submitted twice #12936 <-> #19050 Similarity: 80% Reason: Same code: 10 lines overlap ← Same Telegram thread_id fix, one is literally "v2" #15512 <-> #18994 Similarity: 67% Reason: Same code: 10 lines overlap ← Both normalize Brave search language codes #14182 <-> #15051 Similarity: 75% Reason: Same code: 2403 lines overlap ← Two Zulip implementations duplicating core logic ``` Precision by similarity band, verified by manual sampling: | Band | Pairs | Precision | Notes | |------|-------|-----------|-------| | 100% | 221 | ~75% | High code overlap but tiny sets — watch for short boilerplate | | 80–90% | 318 | ~95% | Best signal — issue refs + code together | | 70–79% | 283 | ~90% | Strong structural duplicates | | 65–69% | 200 | ~70% | Noisier; raise `--threshold 0.75` to cut this band | | **Overall** | **1022** | **~84%** | | For large backlogs, `--threshold 0.75` drops to ~550 pairs at ~92% precision. `--since 30d` gives 52 actionable pairs at ~88% — the recommended default. ## More Examples **VSCode** — Found PRs touching same files for same feature: ``` #295823 <-> #295822 Similarity: 77% Reason: Same files: chatModel.ts, chatForkActions.ts Both: "Use metadata flag for fork detection" ``` **Next.js** — Found related test updates: ``` #90121 <-> #90120 Similarity: 86% Reason: Same files: test/ ``` ## Install ```bash pip install bepo ``` Requires [GitHub CLI](https://cli.github.com/) (`gh`) to be installed and authenticated. ## Usage ```bash # Check a repo for duplicate PRs bepo check --repo owner/repo # Check recent PRs (recommended — avoids stale noise) bepo check --repo owner/repo --since 30d # Adjust sensitivity (default: 0.65, higher = stricter) bepo check --repo owner/repo --threshold 0.7 # JSON output for CI bepo check --repo owner/repo --json ``` ## How It Works bepo fingerprints each PR by extracting: | Signal | Weight | What it catches | |--------|--------|-----------------| | Same issue ref (#123) | 10.0 | Definite duplicate | | Same code changes (IDF-weighted) | 8.0 | Rare lines weighted more than common boilerplate | | Same files touched | 6.0 | PRs modifying same code | | Same feature domain | 3.0 | auth, messaging, database, etc. | | Same imports | 1.0 | Similar dependencies | Then computes pairwise Jaccard similarity. **That's it.** No embeddings, no LLM calls. Just: - Parse `+++ b/path` from diffs - Regex for `#\d+` issue refs - Compare actual code changes - Set intersection for similarity **Cross-component filtering** suppresses boilerplate FPs in integration/plugin monorepos (e.g. Home Assistant, VSCode extensions). Two unrelated integrations sharing scaffold code (`config_flow.py`, `manifest.json`) are filtered out when each PR is concentrated in a different component subtree. Pairs sharing a GitHub issue ref always bypass this filter. ~2,000 lines of Python. ## As a Library ```python from bepo import fingerprint_pr, find_duplicates # Fingerprint PRs fp1 = fingerprint_pr("#123", diff1, title="Fix auth", body="Fixes #456") fp2 = fingerprint_pr("#124", diff2, title="Auth fix", body="Fixes #456") # Find duplicates dups = find_duplicates([fp1, fp2], threshold=0.65) for d in dups: print(f"{d.pr_a} ↔ {d.pr_b}: {d.similarity:.0%}") print(f" Shared issues: {d.shared_issues}") print(f" Shared files: {d.shared_files}") ``` ## GitHub Action Add to your repo to automatically detect duplicate PRs and post a warning comment: ```yaml name: PR Duplicate Check on: [pull_request] jobs: check-duplicates: runs-on: ubuntu-latest steps: - uses: aardpark/bepo@v1 with: github-token: ${{ secrets.GITHUB_TOKEN }} threshold: '0.65' # optional, default 0.65 ``` When a PR is opened that looks like a duplicate, bepo posts a comment: > ## ⚠️ Potential Duplicate PRs Detected > > This PR may be similar to existing open PRs: > > | PR | Similarity | Reason | > |---|---|---| > | [#123](link) | 85% | Both fix #456 | > | [#124](link) | 71% | Same code: 10 lines overlap | > > --- > *Detected by [bepo](https://github.com/aardpark/bepo)* ### Action Inputs | Input | Description | Default | |-------|-------------|---------| | `github-token` | GitHub token for API access | `${{ github.token }}` | | `threshold` | Similarity threshold (0.0-1.0) | `0.65` | | `limit` | Max PRs to compare against | `50` | | `comment` | Post comment on PR | `true` | ### Action Outputs | Output | Description | |--------|-------------| | `has_duplicates` | `true` if duplicates found | | `match_count` | Number of matches | | `matches` | JSON array of matches | ## Why This Works Duplicates share obvious signals: - **Same code** = Identical changes (639 shared lines caught SoundChain duplicates) - **Same issue ref** = Same bug report (#19843 appeared in 4 Matrix PRs) - **Same files** = Same bug location (100% overlap for Feishu cluster) **IDF weighting** makes rare lines matter more than common boilerplate. A shared `startupGraceMs = 5000` is a stronger signal than a shared `return null`. Code overlap and issue refs catch most duplicates. Simple works. ## Origin Story This tool was vibe-coded in a single session with Claude. We tried a few approaches and kept finding that simpler signals outperformed fancier ones. File overlap and issue refs catch most duplicates. Sometimes the obvious solution is the right one. ## License MIT
text/markdown
Andrew Park
null
null
null
null
cli, detection, duplicate, github, pull-request
[ "Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Pyth...
[]
null
null
>=3.10
[]
[]
[]
[]
[]
[]
[]
[ "Homepage, https://github.com/aardpark/bepo", "Repository, https://github.com/aardpark/bepo" ]
twine/6.2.0 CPython/3.14.2
2026-02-19T12:30:25.443018
bepo-1.3.0.tar.gz
31,137
64/4f/9a7e63e4cdd629925ad313ed6262d6702866ab69b825466c68bdfe59e253/bepo-1.3.0.tar.gz
source
sdist
null
false
84d1db3ca5cdd1b8315b4e7dff2e83de
1bb9e9f2efd41a3f97405cf2f82e80b64781125f9ed17bffe90914eea2a7d24f
644f9a7e63e4cdd629925ad313ed6262d6702866ab69b825466c68bdfe59e253
MIT
[ "LICENSE" ]
263
2.4
mex-common
1.16.0
Common library for MEx python projects.
# MEx common Common library for MEx python projects. [![cookiecutter](https://github.com/robert-koch-institut/mex-common/actions/workflows/cookiecutter.yml/badge.svg)](https://github.com/robert-koch-institut/mex-template) [![cve-scan](https://github.com/robert-koch-institut/mex-common/actions/workflows/cve-scan.yml/badge.svg)](https://github.com/robert-koch-institut/mex-common/actions/workflows/cve-scan.yml) [![documentation](https://github.com/robert-koch-institut/mex-common/actions/workflows/documentation.yml/badge.svg)](https://robert-koch-institut.github.io/mex-common) [![linting](https://github.com/robert-koch-institut/mex-common/actions/workflows/linting.yml/badge.svg)](https://github.com/robert-koch-institut/mex-common/actions/workflows/linting.yml) [![open-code](https://github.com/robert-koch-institut/mex-common/actions/workflows/open-code.yml/badge.svg)](https://gitlab.opencode.de/robert-koch-institut/mex/mex-common) [![testing](https://github.com/robert-koch-institut/mex-common/actions/workflows/testing.yml/badge.svg)](https://github.com/robert-koch-institut/mex-common/actions/workflows/testing.yml) ## Project The Metadata Exchange (MEx) project is committed to improve the retrieval of RKI research data and projects. How? By focusing on metadata: instead of providing the actual research data directly, the MEx metadata catalog captures descriptive information about research data and activities. On this basis, we want to make the data FAIR[^1] so that it can be shared with others. Via MEx, metadata will be made findable, accessible and shareable, as well as available for further research. The goal is to get an overview of what research data is available, understand its context, and know what needs to be considered for subsequent use. RKI cooperated with D4L data4life gGmbH for a pilot phase where the vision of a FAIR metadata catalog was explored and concepts and prototypes were developed. The partnership has ended with the successful conclusion of the pilot phase. After an internal launch, the metadata will also be made publicly available and thus be available to external researchers as well as the interested (professional) public to find research data from the RKI. For further details, please consult our [project page](https://www.rki.de/DE/Aktuelles/Publikationen/Forschungsdaten/MEx/metadata-exchange-plattform-mex-node.html). [^1]: FAIR is referencing the so-called [FAIR data principles](https://www.go-fair.org/fair-principles/) – guidelines to make data Findable, Accessible, Interoperable and Reusable. **Contact** \ For more information, please feel free to email us at [mex@rki.de](mailto:mex@rki.de). ### Publisher **Robert Koch-Institut** \ Nordufer 20 \ 13353 Berlin \ Germany ## Package The `mex-common` library is a software development toolkit that is used by multiple components within the MEx project. It contains utilities for building pipelines like a common commandline interface, logging and configuration setup. It also provides common auxiliary connectors that can be used to fetch data from external services and a re-usable implementation of the MEx metadata schema as pydantic models. ## License This package is licensed under the [MIT license](/LICENSE). All other software components of the MEx project are open-sourced under the same license as well. ## Development ### Installation - install python on your system - on unix, run `make install` - on windows, run `.\mex.bat install` ### Linting and testing - run all linters with `make lint` or `.\mex.bat lint` - run unit and integration tests with `make test` or `.\mex.bat test` - run just the unit tests with `make unit` or `.\mex.bat unit` ### Updating dependencies - update boilerplate files with `cruft update` - update global requirements in `requirements.txt` manually - update git hooks with `pre-commit autoupdate` - update package dependencies using `uv sync --upgrade` - update github actions in `.github/workflows/*.yml` manually ### Creating release - run `mex release RULE` to release a new version where RULE determines which part of the version to update and is one of `major`, `minor`, `patch`.
text/markdown
null
MEx Team <mex@rki.de>
null
null
MIT License Copyright (c) 2026 Robert Koch-Institut Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
null
[]
[]
null
null
<3.15,>=3.11
[]
[]
[]
[ "backoff<3,>=2", "click<9,>=8", "langdetect<2,>=1", "ldap3<3,>=2", "mex-model<4.11,>=4.10", "networkx>=3", "numpy<3,>=2", "pandas<4,>=3", "pyarrow<24,>=23", "pydantic-settings<2.13,>=2.12", "pydantic<3,>=2", "pytz>=2025", "requests<3,>=2", "tabulate>=0.9" ]
[]
[]
[]
[ "Repository, https://github.com/robert-koch-institut/mex-common" ]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:29:29.638133
mex_common-1.16.0.tar.gz
89,482
4a/62/8cc0da0a4940f09105897a71edc780d6f0947396e71a4dfe8d1214eec0a9/mex_common-1.16.0.tar.gz
source
sdist
null
false
fe94cb1959d3b8984328a1873ee07f91
4e89e34b206bfb42a1c47092791d430e3b48131a281e2bf86b306698e224a76b
4a628cc0da0a4940f09105897a71edc780d6f0947396e71a4dfe8d1214eec0a9
null
[ "AUTHORS", "LICENSE" ]
339
2.4
smev-agent-client
0.5.1
Клиент для взаимодействия с Агентом ПОДД СМЭВ
# Клиент для взаимодействия со СМЭВ3 посредством Адаптера ## Подключение settings: INSTALLED_APPS = [ 'smev_agent_client', ] apps: from django.apps import AppConfig as AppConfigBase class AppConfig(AppConfigBase): name = __package__ def __setup_agent_client(self): import smev_agent_client smev_agent_client.set_config( smev_agent_client.configuration.Config( agent_url='http://localhost:8090', system_mnemonics='MNSV03', timeout=1, request_retries=1, ) ) def ready(self): super().ready() self.__setup_agent_client() ## Эмуляция Заменить используемый интерфейс на эмулирующий запросы: smev_agent_client.set_config( ..., smev_agent_client.configuration.Config( interface=( 'smev_agent_client.contrib.my_edu.interfaces.rest' '.OpenAPIInterfaceEmulation' ) ) ) ## Запуск тестов $ tox ## API ### Передача сообщения from smev_agent_client.adapters import adapter from smev_agent_client.interfaces import OpenAPIRequest class Request(OpenAPIRequest): def get_url(self): return 'http://localhost:8090/MNSV03/myedu/api/edu-upload/v1/multipart/csv' def get_method(self): return 'post' def get_files(self) -> List[str]: return [ Path('files/myedu_schools.csv').as_posix() ] result = adapter.send(Request())
text/markdown
null
BARS Group <education_dev@bars.group>
null
null
null
null
[ "Development Status :: 5 - Production/Stable", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Framework :: Django :: 3.1", "Framework :: Djang...
[]
null
null
>=3.9
[]
[]
[]
[ "pydantic<2.0,>=1.10.17", "requests<3,>=1.1.0", "Django<5.0,>=3.1", "openapi-core==0.16.5", "isort==5.12.0; extra == \"dev\"", "ruff==0.12.1; extra == \"dev\"", "flake8<7,>=4.0.1; extra == \"dev\"", "pytest<8,>=3.2.5; extra == \"dev\"", "pytest-cov<5; extra == \"dev\"", "sphinx<7.5,>=7; extra == \...
[]
[]
[]
[ "Homepage, https://stash.bars-open.ru/projects/EDUSMEV/repos/smev-agent-client/browse", "Repository, https://stash.bars-open.ru/scm/edusmev/smev-agent-client.git" ]
twine/6.1.0 CPython/3.9.23
2026-02-19T12:29:03.969163
smev_agent_client-0.5.1-py3-none-any.whl
18,314
3b/43/25666898798d2f4a904ac3495460f30e7d02418aa5dab1c86051973881ef/smev_agent_client-0.5.1-py3-none-any.whl
py3
bdist_wheel
null
false
fb83faadd733afcfe7172b7b4a0d060b
da5968ed2d1f91ea0526a015ab40afb8c8b0ff86ba9fe40f161b3a7c129bda2d
3b4325666898798d2f4a904ac3495460f30e7d02418aa5dab1c86051973881ef
null
[ "LICENSE" ]
101
2.4
iwpc
0.8.0
An implementation of the divergence framework as described here https://arxiv.org/abs/2405.06397 and much more...
# IWPC # This package implements the methods described in the research paper https://arxiv.org/abs/2405.06397 for estimating a lower bound on the divergence between any two distributions, p and q, using samples from each distribution. Install using `pip install iwpc` Please see the package [README](https://bitbucket.org/jjhw3/divergences/src/main/) on bitbucket for more information and some examples.
text/markdown
null
"Jeremy J. H. Wilkinson" <jero.wilkinson@gmail.com>
null
null
null
null
[ "Development Status :: 3 - Alpha", "Programming Language :: Python :: 3", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent" ]
[]
null
null
>=3.9
[]
[]
[]
[ "numpy", "torch", "lightning", "matplotlib", "pandas", "scikit-learn", "tensorboard", "seaborn", "bokeh" ]
[]
[]
[]
[ "Homepage, https://bitbucket.org/jjhw3/divergences/src/main/" ]
twine/6.1.0 CPython/3.11.9
2026-02-19T12:28:10.589356
iwpc-0.8.0.tar.gz
83,244
96/80/910908e7dc5cf6b440ecf319b1e5011c99dca7b3668a2024ee1b656d6732/iwpc-0.8.0.tar.gz
source
sdist
null
false
e2e153646ba88af2a905f466e448a8e3
9678973afe67b9ac4d07a1fe441b104d96dbd8434de635e7007c58e94c522b6e
9680910908e7dc5cf6b440ecf319b1e5011c99dca7b3668a2024ee1b656d6732
null
[ "LICENSE" ]
273
2.1
odoo-addon-hr-shift
18.0.1.0.2
Define shifts for employees
.. image:: https://odoo-community.org/readme-banner-image :target: https://odoo-community.org/get-involved?utm_source=readme :alt: Odoo Community Association ================ Employees Shifts ================ .. !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! !! This file is generated by oca-gen-addon-readme !! !! changes will be overwritten. !! !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! !! source digest: sha256:562392450695c6fa5f16c1720b7c6b5dc5f0fb2dda5232242338b27cd2140577 !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! .. |badge1| image:: https://img.shields.io/badge/maturity-Beta-yellow.png :target: https://odoo-community.org/page/development-status :alt: Beta .. |badge2| image:: https://img.shields.io/badge/license-AGPL--3-blue.png :target: http://www.gnu.org/licenses/agpl-3.0-standalone.html :alt: License: AGPL-3 .. |badge3| image:: https://img.shields.io/badge/github-OCA%2Fshift--planning-lightgray.png?logo=github :target: https://github.com/OCA/shift-planning/tree/18.0/hr_shift :alt: OCA/shift-planning .. |badge4| image:: https://img.shields.io/badge/weblate-Translate%20me-F47D42.png :target: https://translation.odoo-community.org/projects/shift-planning-18-0/shift-planning-18-0-hr_shift :alt: Translate me on Weblate .. |badge5| image:: https://img.shields.io/badge/runboat-Try%20me-875A7B.png :target: https://runboat.odoo-community.org/builds?repo=OCA/shift-planning&target_branch=18.0 :alt: Try me on Runboat |badge1| |badge2| |badge3| |badge4| |badge5| Shifts to employees with variable work schedules. **Table of contents** .. contents:: :local: Configuration ============= In order to configure and create shift plannings you'll need to be in the **Shift Manager** security group. Creating shifts --------------- Go to *Shifts > Shifts* and create a new one. Define a name, a color (it will be used in the shifts assignment cards), a start and end time, a time zone and week days span. Create as many as you need to. Setting employees with variable shifts. --------------------------------------- Go to *Employees > Employees* and in the tab *Work information* go to the *Schedule* section. There you can set the **Shift planning** checkbox if this employee should have a shift generated automatically in the weekly plannings. Setting the default work week ----------------------------- Go to *Settings > Employees* and in the *Work organization* section you can define the default working week for your company. By default it goes from Monday to Friday. Usage ===== After configuring the application we can start making plannings. To create a new one: 1. Go to *Shifts > Plannings* and click on *Create*. 2. Set the year and week number for the planning and click *Save*. 3. Now click on *Generate* to pre-create the shifts assignments for your employees. You can start assigning shifts click on the *Shifts* smart button where you'll go to a kanban view with a card per employee that you can drag into the corresponding shift. Once you do it, you'll the color of the week days in the card changes to the color of the shift assigned. |Drag to assign| Now if you want to assign a different shift for a specific day of that week to that employee, you can do so clicking on **Shift details**. In the detailed kanban view drag the days to their corresponding shifts. |Changing specific days| Going back to the general assignment screen you'll see the difference in the days list colors for the employee's card. Every day is clickable and it will pop up the shift details for that specific day. |Card with different shifts| Detecting employees issues -------------------------- An employee could be on leave for one or serveral days of a planning week. In that case when an assignment is made for that employee the overlapping days will be flagged as unavailable and no shift will be assigned. We can detect those issues from the general plannings view in *Shift > Plannings*. |Mark as reviewed| To set the issue as reviewed we can click on the checkbox of the employee's assignment card. It won't be counted on the issues summary when is checked. |image1| Generate planning from another one ---------------------------------- We can generate plannings from other planning so we can copy the shifts assigments. To do so you can either click on **Generate planning** from the general plannings view in *Shifts > Plannings* or click on **Copy to new planning** from the origin planning form. In both cases a wizard will open where you can choose to which week will the new planning correspond to and from which planning we'll be copying the assignations. Regenerate shifts. ------------------ We can reset the assignments from the planning form clicking on *Regenerate shifts*. My shifts --------- All the internal users can view their assigned shifts going to *Shifts > My shifts*. .. |Drag to assign| image:: https://raw.githubusercontent.com/OCA/shift-planning/18.0/hr_shift/static/description/assignment_dragging.gif .. |Changing specific days| image:: https://raw.githubusercontent.com/OCA/shift-planning/18.0/hr_shift/static/description/assignment_details_dragging.gif .. |Card with different shifts| image:: https://raw.githubusercontent.com/OCA/shift-planning/18.0/hr_shift/static/description/week_days_colors.png .. |Mark as reviewed| image:: https://raw.githubusercontent.com/OCA/shift-planning/18.0/hr_shift/static/description/planning_card.png .. |image1| image:: https://raw.githubusercontent.com/OCA/shift-planning/18.0/hr_shift/static/description/reviewed_checkbox.png Known issues / Roadmap ====================== - We can use the *Reviewed* field for more purposes, like setting the planning state when all the shifts are reviewed. - Support working pauses. Bug Tracker =========== Bugs are tracked on `GitHub Issues <https://github.com/OCA/shift-planning/issues>`_. In case of trouble, please check there if your issue has already been reported. If you spotted it first, help us to smash it by providing a detailed and welcomed `feedback <https://github.com/OCA/shift-planning/issues/new?body=module:%20hr_shift%0Aversion:%2018.0%0A%0A**Steps%20to%20reproduce**%0A-%20...%0A%0A**Current%20behavior**%0A%0A**Expected%20behavior**>`_. Do not contact contributors directly about support or help with technical issues. Credits ======= Authors ------- * Tecnativa Contributors ------------ - `Tecnativa <https://tecnativa.com>`__: - David Vidal - Pedro M. Baeza - `Tesseratech <https://www.tesseratech.es>`__: - Abraham Anes - `Grupo Isonor <https://www.grupoisonor.es>`__: - David Palanca Other credits ------------- `Sun Moon icon <https://lucide.dev/icons/sun-moon>`__ by Lucide Maintainers ----------- This module is maintained by the OCA. .. image:: https://odoo-community.org/logo.png :alt: Odoo Community Association :target: https://odoo-community.org OCA, or the Odoo Community Association, is a nonprofit organization whose mission is to support the collaborative development of Odoo features and promote its widespread use. This module is part of the `OCA/shift-planning <https://github.com/OCA/shift-planning/tree/18.0/hr_shift>`_ project on GitHub. You are welcome to contribute. To learn how please visit https://odoo-community.org/page/Contribute.
text/x-rst
Tecnativa, Odoo Community Association (OCA)
support@odoo-community.org
null
null
AGPL-3
null
[ "Programming Language :: Python", "Framework :: Odoo", "Framework :: Odoo :: 18.0", "License :: OSI Approved :: GNU Affero General Public License v3" ]
[]
https://github.com/OCA/shift-planning
null
>=3.10
[]
[]
[]
[ "odoo==18.0.*" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.12.3
2026-02-19T12:27:50.784048
odoo_addon_hr_shift-18.0.1.0.2-py3-none-any.whl
487,301
b0/73/b5c57a206092a891a950d51ed9f6e168f553a13c72b233de699c31bc2e59/odoo_addon_hr_shift-18.0.1.0.2-py3-none-any.whl
py3
bdist_wheel
null
false
27c0760e58b0aa8d1294426837fd1a0f
6e98bf6b64e62c4295448e2bddc608ebb1f542946a4b97c3d9240d5c10e76fc3
b073b5c57a206092a891a950d51ed9f6e168f553a13c72b233de699c31bc2e59
null
[]
95
2.4
aip-proxy
0.1.0
AIP Proxy — Token compression proxy for LLM APIs. Reduce costs 15-40% on any AI IDE (Antigravity, Cursor, VS Code, etc.)
# AIP Proxy **Token compression proxy for LLM APIs.** Reduce your AI coding costs by 15-40% without losing quality. AIP Proxy sits between your AI IDE (Antigravity, Cursor, VS Code, etc.) and the LLM API, compressing prompts on the fly before they reach the model. ## How it works ``` Your IDE ──> AIP Proxy (localhost:8090) ──> OpenAI / Gemini / Claude API │ ├─ Whitespace normalization ├─ Code comment removal ├─ Block deduplication └─ Pattern abbreviation ``` **4 compression passes**, configurable by level: | Level | Passes | Typical savings | |-------|--------|-----------------| | 0 | None (passthrough) | 0% | | 1 | Whitespace normalization | 5-10% | | 2 | + Code compression + deduplication | 15-25% | | 3 | + Pattern abbreviation | 25-40% | ## Install ```bash pip install aip-proxy ``` ## Quick start ```bash # Start proxy targeting OpenAI aip-proxy start --target https://api.openai.com/v1 --port 8090 # Or targeting Google Gemini aip-proxy start --target https://generativelanguage.googleapis.com --port 8090 # Or any OpenAI-compatible API aip-proxy start --target https://api.anthropic.com --port 8090 ``` Then change your IDE's API endpoint to `http://localhost:8090/v1`. ## Usage with Antigravity 1. Install: `pip install aip-proxy` 2. Start: `aip-proxy start --target https://generativelanguage.googleapis.com --port 8090` 3. In Antigravity settings, set API endpoint to `http://localhost:8090` 4. Done — you'll see savings in the proxy stats ## Usage with Cursor / VS Code 1. Install: `pip install aip-proxy` 2. Start: `aip-proxy start --target https://api.openai.com/v1 --port 8090` 3. In your IDE settings, change the API base URL to `http://localhost:8090/v1` 4. Keep your API key as usual — the proxy forwards it transparently ## Options ```bash aip-proxy start --help Options: --target, -t Target API URL (required) --port, -p Port to listen on (default: 8090) --host Host to bind (default: 127.0.0.1) --level, -l Compression: 0=off, 1=light, 2=balanced, 3=aggressive (default: 2) --no-cache Disable response caching --cache-ttl Cache TTL in seconds (default: 300) ``` ## Endpoints | Endpoint | Description | |----------|-------------| | `GET /health` | Proxy status and basic stats | | `GET /stats` | Detailed compression and cache statistics | | `* /{path}` | Proxied to target API | ## Python API ```python from aip_proxy import TokenCompressor tc = TokenCompressor(level=2) messages = [ {"role": "user", "content": "your long prompt here..."} ] compressed = tc.compress_messages(messages) print(tc.get_savings()) # {'original_chars': 1500, 'compressed_chars': 1100, 'saved_chars': 400, 'savings_pct': 26.7, 'calls': 1} ``` ## How does it save money? LLM APIs charge per token. A typical coding session sends thousands of tokens in context — much of it is: - Redundant whitespace and blank lines - Comments in code blocks (the model doesn't need them) - Repeated code blocks across messages - Verbose filler phrases AIP Proxy removes this noise while preserving the semantic content the model needs. ## License MIT
text/markdown
Carmen Esteban
null
null
null
MIT
llm-proxy, token-compression, ai-cost-reduction, openai-proxy, antigravity, cursor, vscode, aip-engine, prompt-optimization
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Py...
[]
null
null
>=3.9
[]
[]
[]
[ "fastapi>=0.100", "uvicorn>=0.20", "httpx>=0.24", "tiktoken>=0.5" ]
[]
[]
[]
[ "Homepage, https://github.com/iafiscal1212/aip-proxy", "Repository, https://github.com/iafiscal1212/aip-proxy" ]
twine/6.2.0 CPython/3.12.3
2026-02-19T12:27:26.411582
aip_proxy-0.1.0.tar.gz
12,508
71/60/783f4a03d127c7f6843ff25e873661d72940b0f093f1713ceec34f2591f9/aip_proxy-0.1.0.tar.gz
source
sdist
null
false
1508664509251fb7c589b4e9b06b0672
60e1479feaf561c0caf204b1d8feee3ae6fd11ae03eb7b77f0423dd64f18436f
7160783f4a03d127c7f6843ff25e873661d72940b0f093f1713ceec34f2591f9
null
[ "LICENSE" ]
301
2.4
albert
1.18.0
The official Python SDK for the Albert Invent platform.
# Albert Python SDK [![PyPI version](https://img.shields.io/pypi/v/albert.svg)](https://pypi.org/project/albert/) [![Python versions](https://img.shields.io/badge/python-3.10%20%7C%203.11%20%7C%203.12%20%7C%203.13-blue.svg)](https://www.python.org/) [![Downloads per month](https://img.shields.io/pypi/dm/albert.svg)](https://pypi.org/project/albert/) Albert Python is the official Albert Invent Software Development Kit (SDK) for Python that provides a comprehensive and easy-to-use interface for interacting with the Albert Platform. The SDK allows Python developers to write software that interacts with various platform resources, such as inventories, projects, companies, tags, and many more. You can find the latest, most up-to-date documentation on the supported resources and usage patterns [here](https://docs.developer.albertinvent.com/albert-python). ## Installation `pip install albert` This installs the latest stable release from [PyPI](https://pypi.org/project/albert/). ### Contribution For developers, please see the [contributing guide](CONTRIBUTING.md), which includes setup instructions, testing, and linting guidelines. ## Quick Start ```python from albert import Albert client = Albert.from_client_credentials( base_url="https://app.albertinvent.com", client_id=YOUR_CLIENT_ID, client_secret=YOUR_CLIENT_SECRET ) projects = client.projects.get_all() ``` ## Documentation [Full Documentation can be found here](https://docs.developer.albertinvent.com/albert-python/latest/)
text/markdown
null
Albert Invent <support@albertinvent.com>
null
null
Apache License Version 2.0, January 2004 http://www.apache.org/licenses/ TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION 1. Definitions. "License" shall mean the terms and conditions for use, reproduction, and distribution as defined by Sections 1 through 9 of this document. "Licensor" shall mean the copyright owner or entity authorized by the copyright owner that is granting the License. "Legal Entity" shall mean the union of the acting entity and all other entities that control, are controlled by, or are under common control with that entity. For the purposes of this definition, "control" means (i) the power, direct or indirect, to cause the direction or management of such entity, whether by contract or otherwise, or (ii) ownership of fifty percent (50%) or more of the outstanding shares, or (iii) beneficial ownership of such entity. "You" (or "Your") shall mean an individual or Legal Entity exercising permissions granted by this License. "Source" form shall mean the preferred form for making modifications, including but not limited to software source code, documentation source, and configuration files. "Object" form shall mean any form resulting from mechanical transformation or translation of a Source form, including but not limited to compiled object code, generated documentation, and conversions to other media types. "Work" shall mean the work of authorship, whether in Source or Object form, made available under the License, as indicated by a copyright notice that is included in or attached to the work (an example is provided in the Appendix below). "Derivative Works" shall mean any work, whether in Source or Object form, that is based on (or derived from) the Work and for which the editorial revisions, annotations, elaborations, or other modifications represent, as a whole, an original work of authorship. For the purposes of this License, Derivative Works shall not include works that remain separable from, or merely link (or bind by name) to the interfaces of, the Work and Derivative Works thereof. "Contribution" shall mean any work of authorship, including the original version of the Work and any modifications or additions to that Work or Derivative Works thereof, that is intentionally submitted to Licensor for inclusion in the Work by the copyright owner or by an individual or Legal Entity authorized to submit on behalf of the copyright owner. For the purposes of this definition, "submitted" means any form of electronic, verbal, or written communication sent to the Licensor or its representatives, including but not limited to communication on electronic mailing lists, source code control systems, and issue tracking systems that are managed by, or on behalf of, the Licensor for the purpose of discussing and improving the Work, but excluding communication that is conspicuously marked or otherwise designated in writing by the copyright owner as "Not a Contribution." "Contributor" shall mean Licensor and any individual or Legal Entity on behalf of whom a Contribution has been received by Licensor and subsequently incorporated within the Work. 2. Grant of Copyright License. Subject to the terms and conditions of this License, each Contributor hereby grants to You a perpetual, worldwide, non-exclusive, no-charge, royalty-free, irrevocable copyright license to reproduce, prepare Derivative Works of, publicly display, publicly perform, sublicense, and distribute the Work and such Derivative Works in Source or Object form. 3. Grant of Patent License. Subject to the terms and conditions of this License, each Contributor hereby grants to You a perpetual, worldwide, non-exclusive, no-charge, royalty-free, irrevocable (except as stated in this section) patent license to make, have made, use, offer to sell, sell, import, and otherwise transfer the Work, where such license applies only to those patent claims licensable by such Contributor that are necessarily infringed by their Contribution(s) alone or by combination of their Contribution(s) with the Work to which such Contribution(s) was submitted. If You institute patent litigation against any entity (including a cross-claim or counterclaim in a lawsuit) alleging that the Work or a Contribution incorporated within the Work constitutes direct or contributory patent infringement, then any patent licenses granted to You under this License for that Work shall terminate as of the date such litigation is filed. 4. Redistribution. You may reproduce and distribute copies of the Work or Derivative Works thereof in any medium, with or without modifications, and in Source or Object form, provided that You meet the following conditions: (a) You must give any other recipients of the Work or Derivative Works a copy of this License; and (b) You must cause any modified files to carry prominent notices stating that You changed the files; and (c) You must retain, in the Source form of any Derivative Works that You distribute, all copyright, patent, trademark, and attribution notices from the Source form of the Work, excluding those notices that do not pertain to any part of the Derivative Works; and (d) If the Work includes a "NOTICE" text file as part of its distribution, then any Derivative Works that You distribute must include a readable copy of the attribution notices contained within such NOTICE file, excluding those notices that do not pertain to any part of the Derivative Works, in at least one of the following places: within a NOTICE text file distributed as part of the Derivative Works; within the Source form or documentation, if provided along with the Derivative Works; or, within a display generated by the Derivative Works, if and wherever such third-party notices normally appear. The contents of the NOTICE file are for informational purposes only and do not modify the License. You may add Your own attribution notices within Derivative Works that You distribute, alongside or as an addendum to the NOTICE text from the Work, provided that such additional attribution notices cannot be construed as modifying the License. You may add Your own copyright statement to Your modifications and may provide additional or different license terms and conditions for use, reproduction, or distribution of Your modifications, or for any such Derivative Works as a whole, provided Your use, reproduction, and distribution of the Work otherwise complies with the conditions stated in this License. 5. Submission of Contributions. Unless You explicitly state otherwise, any Contribution intentionally submitted for inclusion in the Work by You to the Licensor shall be under the terms and conditions of this License, without any additional terms or conditions. Notwithstanding the above, nothing herein shall supersede or modify the terms of any separate license agreement you may have executed with Licensor regarding such Contributions. 6. Trademarks. This License does not grant permission to use the trade names, trademarks, service marks, or product names of the Licensor, except as required for reasonable and customary use in describing the origin of the Work and reproducing the content of the NOTICE file. 7. Disclaimer of Warranty. Unless required by applicable law or agreed to in writing, Licensor provides the Work (and each Contributor provides its Contributions) on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied, including, without limitation, any warranties or conditions of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A PARTICULAR PURPOSE. You are solely responsible for determining the appropriateness of using or redistributing the Work and assume any risks associated with Your exercise of permissions under this License. 8. Limitation of Liability. In no event and under no legal theory, whether in tort (including negligence), contract, or otherwise, unless required by applicable law (such as deliberate and grossly negligent acts) or agreed to in writing, shall any Contributor be liable to You for damages, including any direct, indirect, special, incidental, or consequential damages of any character arising as a result of this License or out of the use or inability to use the Work (including but not limited to damages for loss of goodwill, work stoppage, computer failure or malfunction, or any and all other commercial damages or losses), even if such Contributor has been advised of the possibility of such damages. 9. Accepting Warranty or Additional Liability. While redistributing the Work or Derivative Works thereof, You may choose to offer, and charge a fee for, acceptance of support, warranty, indemnity, or other liability obligations and/or rights consistent with this License. However, in accepting such obligations, You may act only on Your own behalf and on Your sole responsibility, not on behalf of any other Contributor, and only if You agree to indemnify, defend, and hold each Contributor harmless for any liability incurred by, or claims asserted against, such Contributor by reason of your accepting any such warranty or additional liability. END OF TERMS AND CONDITIONS APPENDIX: How to apply the Apache License to your work. To apply the Apache License to your work, attach the following boilerplate notice, with the fields enclosed by brackets "[]" replaced with your own identifying information. (Don't include the brackets!) The text should be enclosed in the appropriate comment syntax for the file format. We also recommend that a file or class name and description of purpose be included on the same "printed page" as the copyright notice for easier identification within third-party archives. Copyright [2025] [Albert Invent] Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.
API, Albert, Python, SDK
[ "Development Status :: 3 - Alpha", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3" ]
[]
null
null
>=3.10
[]
[]
[]
[ "pandas<3,>=2.2.2", "pydantic[email]<3,>=2.8.2", "pyjwt<3,>=2.10.0", "requests<3,>=2.32.3", "tenacity>=8.2.3" ]
[]
[]
[]
[ "Homepage, https://www.albertinvent.com/", "Documentation, https://docs.developer.albertinvent.com/albert-python", "Repository, https://github.com/albert-labs/albert-python", "Issues, https://github.com/albert-labs/albert-python/issues" ]
uv/0.10.4 {"installer":{"name":"uv","version":"0.10.4","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"22.04","id":"jammy","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
2026-02-19T12:27:22.672315
albert-1.18.0.tar.gz
646,533
40/0e/cf1032f42ee154380512685e3f12c2e87390669b68135e3a82a079ebaa39/albert-1.18.0.tar.gz
source
sdist
null
false
be53237b6a669a95e0929873c500ee4b
c3da712917b8fcbd1dfff1e97860fe6722cb445372597eee794616cded481a4a
400ecf1032f42ee154380512685e3f12c2e87390669b68135e3a82a079ebaa39
null
[ "LICENSE" ]
293
2.4
sphinxcontrib-svgbob
0.3.2
A Sphinx extension to convert ASCII diagrams to SVGs with Svgbob.
# `sphinxcontrib-svgbob` [![Stars](https://img.shields.io/github/stars/sphinx-contrib/svgbob.svg?style=social&maxAge=3600&label=Star)](https://github.com/sphinx-contrib/svgbob/stargazers) *A Sphinx extension to render ASCII diagrams into SVG using [Svgbob](https://github.com/ivanceras/svgbob).* [![Actions](https://img.shields.io/github/actions/workflow/status/sphinx-contrib/svgbob/test.yml?style=flat-square&maxAge=600)](https://github.com/sphinx-contrib/svgbob/actions) [![Codecov](https://img.shields.io/codecov/c/gh/sphinx-contrib/svgbob/master.svg?style=flat-square&maxAge=600)](https://codecov.io/gh/sphinx-contrib/svgbob) [![License](https://img.shields.io/badge/license-MIT-blue.svg?style=flat-square&maxAge=2678400)](https://choosealicense.com/licenses/mit/) [![Source](https://img.shields.io/badge/source-GitHub-303030.svg?maxAge=2678400&style=flat-square)](https://github.com/sphinx-contrib/svgbob/) [![PyPI](https://img.shields.io/pypi/v/sphinxcontrib-svgbob.svg?style=flat-square&maxAge=600)](https://pypi.org/project/sphinxcontrib-svgbob) [![Wheel](https://img.shields.io/pypi/wheel/sphinxcontrib-svgbob.svg?style=flat-square&maxAge=2678400)](https://pypi.org/project/sphinxcontrib-svgbob/#files) [![Python Versions](https://img.shields.io/pypi/pyversions/sphinxcontrib-svgbob.svg?style=flat-square&maxAge=600)](https://pypi.org/project/sphinxcontrib-svgbob/#files) [![PyPI - Implementation](https://img.shields.io/pypi/implementation/sphinxcontrib-svgbob.svg?style=flat-square&maxAge=600)](https://pypi.org/project/sphinxcontrib-svgbob/#files) [![Changelog](https://img.shields.io/badge/keep%20a-changelog-8A0707.svg?maxAge=2678400&style=flat-square)](https://github.com/sphinx-contrib/svgbob/blob/master/CHANGELOG.md) [![GitHub issues](https://img.shields.io/github/issues/sphinx-contrib/svgbob.svg?style=flat-square&maxAge=600)](https://github.com/sphinx-contrib/svgbob/issues) [![Downloads](https://img.shields.io/pypi/dm/sphinxcontrib-svgbob?style=flat-square&color=303f9f&maxAge=86400&label=downloads)](https://pepy.tech/project/sphinxcontrib-svgbob) ## 🗺️ Overview Diagrams to be included into Sphinx documentation are commonly described with a dedicated markup language, and converted into an image by Sphinx when the documentation is built. However, this reduces the legibility of the documentation source for readers that are not browsing the HTML version. [Svgbob](https://github.com/ivanceras/svgbob) is a diagramming model implemented in Rust that can convert ASCII diagrams into SVG. Using it allows you to: * Keep a textual version of the diagram in your documentation, so that it remains legible. * Render a nicer version as SVG for HTML or PDF versions of the documentation. This Sphinx extension builds Svgbob statically and lets you use it to render ASCII diagrams within Sphinx documentation. Since it does not require any external dependency, it's also suitable to use on [readthedocs.org](https://readthedocs.org). ## 🔧 Installing `sphinxcontrib-svgbob` can be installed from [PyPI](https://pypi.org/project/sphinxcontrib-svgbob/), which hosts some pre-built CPython wheels for x86-64 Linux and OSX, as well as the code required to compile from source: ```console $ pip install sphinxcontrib-svgbob ``` If a Rust compiler is not available, the `setup.py` script will attempt to install a temporary copy if the package is compiled on a UNIX system. If it doesn't work, see the [documentation on `rust-lang.org`](https://forge.rust-lang.org/other-installation-methods.html) to learn how to install Rust on your machine. Then add this extension to the Sphinx extensions in your `conf.py` file to make the `svgbob` directive available: ```python extensions = [ ..., "sphinxcontrib.svgbob", ] ``` That's it, you're all set! ## 💡 Example Use the `svgbob` directive in a function docstring to show a diagram of what is being computed: ```python def hamming(x, y): """Compute the Hamming distance between two strings. Hamming distance between two strings of equal length is the number of positions at which the corresponding symbols are different. For instance, Hamming distance for a 3-bit string can be computed visually using a 3-bit binary cube: .. svgbob:: :align: center 110 111 *-----------* /| /| / | 011 / | 010 *--+--------* | | | 100 | | | *--------+--* 101 | / | / |/ |/ 000 *-----------* 001 The minimum distance between any two vertices is the Hamming distance between the two bit vectors (e.g. 100→011 has distance 3). """ ``` When Sphinx (and `autodoc`) renders the docstring of this function, you'll get the following HTML page (here shown with the [Sphinx theme for readthedocs.org](https://github.com/readthedocs/sphinx_rtd_theme)): ![example1.html.png](https://raw.githubusercontent.com/sphinx-contrib/svgbob/master/static/example1.html.png) And yet, the `help(hamming)` will still look nice and helpful: ![example1.console.png](https://raw.githubusercontent.com/sphinx-contrib/svgbob/master/static/example1.console.png) ## 🔩 Configuration The `svgbob` directive supports the following arguments: - `font-size` (integer): the size of the text to be rendered, defaults to *14*. - `font-family`: the family of the font used to render the text, defaults to *monospace*. - `fill-color` (CSS color): the color to use to fill closed shapes. - `stroke-color` (CSS color): the color to use to paint strokes, defaults to *black*. - `scale` (float): the SVG scale of the figure, defaults to *8.0*. - `align` (CSS align value): the alignment of the resulting image. - `class` (HTML class): an arbitrary class to add to the resulting HTML element. For instance, use the following to use Arial with size 12, to render nicer text in the diagram blocks: ```rst .. svgbob:: :font-family: Arial :font-size: 12 +-------+ +--------+ | Hello |------>| World! | +-------+ +--------+ ``` ![example2.svg](https://raw.githubusercontent.com/sphinx-contrib/svgbob/master/static/example2.svg) ## 💭 Feedback ### ⚠️ Issue Tracker Found a bug ? Have an enhancement request ? Head over to the [GitHub issue tracker](https://github.com/sphinx-contrib/svgbob/issues) if you need to report or ask something. If you are filing in on a bug, please include as much information as you can about the issue, and try to recreate the same bug in a simple, easily reproducible situation. ### 🏗️ Contributing Contributions are more than welcome! See [`CONTRIBUTING.md`](https://github.com/sphinx-contrib/svgbob/blob/master/CONTRIBUTING.md) for more details. ## 📚 Alternatives * [`sphinxcontrib-kroki`](https://github.com/sphinx-contrib/kroki/) also lets you use Svgbob to convert ASCII diagrams, but it queries the [kroki.io](https://kroki.io/) website to do so, and does not support the new options from Svgbob v0.5. * [`sphinxcontrib-aafig`](https://github.com/sphinx-contrib/aafig) uses the [`aafigure`](https://launchpad.net/aafigure) binary to convert ASCII diagrams. ## 🔨 Credits `sphinxcontrib-svgbob` is developped and maintained by: - [Martin Larralde](https://github.com/althonos) The structure of this repository was adapted from the aforementioned `sphinxcontrib-kroki` repository, as I had no experience setting up a Sphinx extension otherwise. ## ⚖️ License This library is provided under the [MIT License](https://choosealicense.com/licenses/mit/).
text/markdown; charset=UTF-8; variant=GFM
null
Martin Larralde <martin.larralde@embl.de>
null
null
null
sphinx, documentation, svg, diagram, ascii
[ "Development Status :: 4 - Beta", "Framework :: Sphinx :: Extension", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", ...
[]
https://pypi.org/projects/sphinxcontrib-svgbob
null
>=3.8
[]
[]
[]
[ "sphinx" ]
[]
[]
[]
[ "Bug Tracker, https://github.com/sphinx-contrib/svgbob/issues", "Changelog, https://github.com/sphinx-contrib/svgbob/blob/master/CHANGELOG.md", "PyPI, https://pypi.org/project/sphinxcontrib-svgbob" ]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:27:07.096529
sphinxcontrib_svgbob-0.3.2.tar.gz
21,593
99/7b/0961a42fa052935cb8320f4e2590d2d435ba8fe53d7b0d019772516e1def/sphinxcontrib_svgbob-0.3.2.tar.gz
source
sdist
null
false
f9ad71edb2505ec5dce26515e5b9d9b9
6b17b278d4a1cb46c85e29b55f75ecd272b4098e7a4e19db3bdab37b426f75cb
997b0961a42fa052935cb8320f4e2590d2d435ba8fe53d7b0d019772516e1def
null
[ "COPYING" ]
1,310
2.1
lib4vex
0.2.2
VEX generator and consumer library
# Lib4VEX Lib4VEX is a library to parse and generate VEX documents. It supports VEX documents created in the [OpenVEX](https://openvex.dev), [CycloneDX](https://www.cyclonedx.org) or [CSAF](https://docs.oasis-open.org/csaf/csaf/v2.0/csaf-v2.0.html) specifications. It has been developed on the assumption that having a generic abstraction of vulnerability regardless of the underlying format will be useful to developers. The following facilities are provided: - Generate OpenVEX, CycloneDX and CSAF VEX documents in JSON format - Parse CycloneDX SBOM in JSON format and extract vulnerability information - Parse OpenVEX and CSAF documents to extract vulnerability information - Generated VEX document can be output to a file or to the console ## Installation To install use the following command: `pip install lib4vex` Alternatively, just clone the repo and install dependencies using the following command: `pip install -U -r requirements.txt` The tool requires Python 3 (3.8+). It is recommended to use a virtual python environment especially if you are using different versions of python. `virtualenv` is a tool for setting up virtual python environments which allows you to have all the dependencies for the tool set up in a single environment, or have different environments set up for testing using different versions of Python. ## API ### Metadata ### Product ### Vulnerability ### Debug Creating the environment variable _**LIB4VEX_DEBUG**_ will result in some additional information being reported when a VEX document is being generated. ## Examples A number of example scripts are included in the _examples_ subdirectory. Examples are provided for CSAF, CycloneDX and OpenVEX scenarios. ## Tutorial A tutorial showing a lifecycle of vulnerabilities is [available](TUTORIAL.md). Whilst the tutorial uses CSAF as the VEX document, equivalent steps can be performed for producing a VEX document using CycloneDX or OpenVEX. ## Implementation Notes The following design decisions have been made in creating and processing VEX files: 1. VEXes should be produced with reference to an SBOM so that only vulnerabilities for components included in the SBOM are included in the VEX document. 2. The VEX document contains all reported vulnerabilities and the respective status. The latest VEX is indicated by the latest timestamp. The previous VEX documents are retained for audit purposes. 3. The VEX document is intended to be used for a single product. ## Future Development 1. Add support for SPDX Security profile when released as part of the SPDX 3.0 release. ## License Licensed under the Apache 2.0 Licence. ## Limitations This library is meant to support software development. The usefulness of the library is dependent on the data which is provided. Unfortunately, the library is unable to determine the validity or completeness of such a VEX file; users of the library and the resulting VEX file are therefore reminded that they should assert the quality of any data which is provided to the library. ## Feedback and Contributions Bugs and feature requests can be made via GitHub Issues.
text/markdown
Anthony Harrison
anthony.p.harrison@gmail.com
Anthony Harrison
anthony.p.harrison@gmail.com
Apache-2.0
security, tools, SBOM, DevSecOps, SPDX, CycloneDX, VEX, CSAF, OpenVEX, library
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Py...
[]
https://github.com/anthonyharrison/lib4vex
null
>=3.7
[]
[]
[]
[ "lib4sbom>=0.8.8", "csaf-tool>=0.3.2", "packageurl-python" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.10.8
2026-02-19T12:27:04.886716
lib4vex-0.2.2-py3-none-any.whl
23,062
ec/f0/d6e2a00173ecebee0a6afc05ac6e8cbd96b1a689bada780dd680dcb550d9/lib4vex-0.2.2-py3-none-any.whl
py3
bdist_wheel
null
false
653805d68fbdffc5ca47ecbea65300dc
c1f54beb25c61e1f59be0367b6e9608a260f7cf49b9d292b3b98cfe032383672
ecf0d6e2a00173ecebee0a6afc05ac6e8cbd96b1a689bada780dd680dcb550d9
null
[]
2,914
2.4
9z
0.1.1
Simple HiHello responder library
# HiHello HiHello
text/markdown
null
Oywe <ncql@proton.me>
null
null
MIT
null
[]
[]
null
null
>=3.8
[]
[]
[]
[]
[]
[]
[]
[]
twine/6.2.0 CPython/3.11.14
2026-02-19T12:27:02.003250
9z-0.1.1.tar.gz
1,443
0f/8d/ec2f85908a6790ee7d72df1226afd79ff1795c32af372195d01a79e2a743/9z-0.1.1.tar.gz
source
sdist
null
false
b3c401a7e0e672ddafc5b88e67fab674
15c2154a8dd1a29c4d6ec69870b2ee1979a0e4b83aa52f243e20558851ec7837
0f8dec2f85908a6790ee7d72df1226afd79ff1795c32af372195d01a79e2a743
null
[]
310
2.4
rpa-quaestvm
1.6.4
Abstrações e utilidades da biblioteca pyautogui e outras ferramentas de RPA
# RPA Quaestvm Abstrações e implementações úteis para RPA ## Funcionalidades ### Classe Logger Abstração do logging padrão do python. Cria uma única instância do logger para cada caminho de arquivo. No construtor, deve-se passar o caminho do arquivo de logs criado, o nome do logger (opcional) e o nível de logs (opcional) ```python def __init__(self, logs_path: str, name: str = 'RPA_Logger', level: str = 'INFO') ``` O método get_logger retorna a instância do objeto de acordo com o valor de logs_path, ou cria uma nova. ### classe Pyautoqstvm Essa clase possui métodos úteis para navegação desktop usando pyautogui, listados abaixo. No construtor, pode-se passar um objeto da classe rpa_quaestvm.Logger e o caminho padrão das imagens de mapemaento: ```python from rpa_quaestvm.logger import Logger from rpa_quaestvm.pyautoqstvm import Pyautoqstvm logger = Logger.get_logger("logs/rpa.log") pyautoqstvm = Pyautoqstvm(logger=logger, pasta_imagens="mapeamento") pyautoqstvm.espera_imagem(...) ``` ```python def espera_imagem( self, imagem: str | list[str], timeout: int = 10, confidence: float = 0.9, raise_errors: bool = True, region: tuple[int, int, int, int] = None, verbose: bool = True, pasta_imagens: str = None, extensao_imagem: str = 'png' ): """Aguarda uma imagem surgir na tela e captura sua posição usando pyautogui Args: imagem (str): nome da imagem a buscar, sem extensão e pasta raiz. o caminho completo será a concatenação dos parâmetros pasta_imagens, imagem e extensao_imagem timeout (int, optional): quantos segundos esperar pela imagem na tela. Defaults to 10. confidence (float, optional): parâmetro de confiança passado para o pyautogui. interfere em quanto a imagem real precisa estar fiel à passada por parâmetro. Defaults to 0.9. raise_errors (bool, optional): se True, lança erro quando não encontra a imagem, se não retorna um booleano. Defaults to True. region (tuple[int, int, int, int], optional): região específica da tela na qual procurar a imagem. Defaults to None. verbose (bool, optional): se deve ou não logar informações usando o logger da biblioteca. Defaults to True. pasta_imagens (str, optional): pasta na qual as imagens de mapeamento são salvas, relativa à raiz. Sobrepõe o valor passado no construtor, se tiver passado. Defaults to "mapeamento". extensao_imagem (str, optional): extensão da imagem, sem o ponto. Defaults to 'png'. Returns: bool: se raise_errors for False e não achar a imagem Box: se encontrar a imagem, retorna o objeto Box do pyautogui """ ``` ```python def get_coordenadas_location( self, location: Box, horizontal: Literal["esquerda", "direita", "centro"] = "centro", vertical: Literal["cima", "baixo", "centro"] = "centro", ): """busca as coordenadas de uma região específica dentro de uma Box A região pode ser uma combinação dos valores horizontal e vertical, sendo retornada a coordenada x e y daquele ponto. Args: location (Box): objeto retornado pelo pyautogui.locateOnScreen ou pelo pyautoqstvm.espera_imagem horizontal (Literal[&quot;esquerda&quot;, &quot;direita&quot;, &quot;centro&quot;], optional): região horizontal. Defaults to "centro". vertical (Literal[&quot;cima&quot;, &quot;baixo&quot;, &quot;centro&quot;], optional): região vertical. Defaults to "centro". Raises: ValueError: se os parâmetros horizontal ou vertical não forem válidos Returns: tuple(int, int): coordenadas x e y do ponto escolhido """ ``` ```python def espera_imagem_e_clica( self, imagem: str | list[str], click: Literal["simples", "duplo", "triplo", "nada", "direito", "rodinha"] = "simples", horizontal: Literal["esquerda", "direita", "centro"] = "centro", vertical: Literal["cima", "baixo", "centro"] = "centro", timeout: int = 10, confidence: float = 0.9, raise_errors: bool = True, region: tuple[int, int, int, int] = None, verbose: bool = True, pasta_imagens: str = None, extensao_imagem: str = 'png' ) -> bool: """ Espera por uma imagem na tela e clica nela. Usa os métodos espera_imagem e get_coordenadas_location Args: imagem (str): O caminho para o arquivo da imagem (.png) a ser localizada. click_type (str): Tipo de clique: "simples" para clique único, "duplo" para clique duplo. "nada" para não clicar. (Padrão: "simples") horizontal (str): Posição horizontal do clique dentro da imagem: "esquerda", "direita" ou "centro". (Padrão: "centro") vertical (str): Posição vertical do clique dentro da imagem: "cima", "baixo" ou "centro". (Padrão: "centro") timeout (int): Tempo máximo em segundos para esperar pela imagem. (Padrão: 10) confidence (float): Nível de confiança para a detecção da imagem (0.0 a 1.0). Valores mais altos são mais estritos. (Padrão: 0.9) raise_errors (bool, optional): se True, lança erro quando não encontra a imagem, se não retorna um booleano. Defaults to True. region (tuple[int, int, int, int], optional): região específica da tela na qual procurar a imagem. Defaults to None. verbose (bool, optional): se deve ou não logar informações usando o logger da biblioteca. Defaults to True. pasta_imagens (str, optional): pasta na qual as imagens de mapeamento são salvas, relativa à raiz. Sobrepõe o valor passado no construtor, se tiver passado. Defaults to "mapeamento". extensao_imagem (str, optional): extensão da imagem, sem o ponto. Defaults to 'png'. Returns: bool: se raise_errors for False e não achar a imagem Box: se encontrar a imagem, retorna o objeto Box do pyautogui """ ``` ```python def limpar_campo( self, campo: str | tuple[int, int], horizontal: str = "direita", vertical: str = "centro", ): """Limpa um campo de texto usando atalhos de teclado (ctrl+a -> delete) Args: campo (str | tuple[int, int]): imagem a clicar (usando espera_imagem_e_clica), ou coordenadas x e y para clicar usando pyautogui.click horizontal (str, optional): região horizontal do clique, se passou uma imagem como campo. Defaults to "direita". vertical (str, optional): região vertical do clique, se passou uma imagem como campo. Defaults to "centro". """ ``` ```python def validar_presenca_imagem( self, imagem: str, clique: Literal[ "simples", "duplo", "triplo", "nada", "direito", "rodinha" ] = "nada", horizontal: Literal["esquerda", "direita", "centro"] = "centro", vertical: Literal["cima", "baixo", "centro"] = "centro", msg_erro: str = "Erro de validação de imagem", deve_existir: bool = True, campo_a_limpar: str | tuple[int, int] = None, timeout: int = 1, on_error: Callable | None = None, confidence: float = 0.9, region: tuple[int, int, int, int] = None, ): """Espera por uma imagem aparecer na tela, podendo realizar callbacks caso a imagem exista ou não Args: imagem (str): nome da imagem a buscar, sem extensão e pasta raiz. o caminho completo será a concatenação dos parâmetros pasta_imagens, imagem e extensao_imagem clique (str): Tipo de clique: "simples" para clique único, "duplo" para clique duplo. "nada" para não clicar. (Padrão: "simples") horizontal (str): Posição horizontal do clique dentro da imagem: "esquerda", "direita" ou "centro". (Padrão: "centro") vertical (str): Posição vertical do clique dentro da imagem: "cima", "baixo" ou "centro". (Padrão: "centro") timeout (int): Tempo máximo em segundos para esperar pela imagem. (Padrão: 10) confidence (float): Nível de confiança para a detecção da imagem (0.0 a 1.0). Valores mais altos são mais estritos. (Padrão: 0.9) region (tuple[int, int, int, int], optional): região específica da tela na qual procurar a imagem. Defaults to None. verbose (bool, optional): se deve ou não logar informações usando o logger da biblioteca. Defaults to True. pasta_imagens (str, optional): pasta na qual as imagens de mapeamento são salvas, relativa à raiz. Defaults to "mapeamento". extensao_imagem (str, optional): extensão da imagem, sem o ponto. Defaults to 'png'. msg_erro (str, optional): mensagem de erro lançada caso a validação falhe. Defaults to "Erro de validação de imagem". deve_existir (bool, optional): se a imagem deve ser encontrada ou não. Se for False, a existência da imagem se torna um erro. Defaults to True. campo_a_limpar (str | tuple[int, int], optional): nome da imagem ou coordenadas x e y passadas para o método limpar_campo caso a validação falhe. Defaults to None. on_error (Callable | None, optional): método chamado em caso de erro, sem parâmetros. Defaults to None. Raises: Exception: caso a validação falhe """ ``` ```python def encerrar_task(self, task: str): """Encerra uma task do sistema usando taskkill Args: task (str): nome da task usada no taskkill """ ``` ```python def abrir_e_focar_aplicacao( self, caminho_executavel: str = None, titulo_janela_re: str = ".*", tempo_limite: int = 30, ): """ Inicia uma aplicação (se o caminho for fornecido) ou foca em uma já existente, aguardando sua janela principal e dando foco a ela, utilizando apenas pyautogui. Args: caminho_executavel (str, opcional): O caminho completo para o executável da aplicação (ex: "C:\\Windows\\notepad.exe"). Se não for fornecido, a função tentará focar em uma janela existente com o título especificado. titulo_janela_re (str): Uma expressão regular para o título da janela principal esperada. Use ".*" para qualquer título (padrão). tempo_limite (int): O tempo máximo em segundos para esperar a janela carregar. Returns: bool: True se a aplicação foi aberta/focada e a janela principal foi focada com sucesso, False caso contrário. """ ```
text/markdown
null
Wellington Velasco <wvelasco@quaestum.com.br>, Indira Lima <indira.lima@quaestum.com.br>
null
null
null
null
[ "Operating System :: OS Independent", "Programming Language :: Python :: 3" ]
[]
null
null
>=3.9
[]
[]
[]
[ "easyocr", "opencv-python", "pillow", "plyer", "psutil", "pyautogui", "pygetwindow", "pyscreeze", "python-dotenv" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.13.3
2026-02-19T12:26:10.790002
rpa_quaestvm-1.6.4.tar.gz
32,074
52/b1/a15617ec28a679a2929fd2c0efdea89a0db7bd74ceb9f0f4624eecbfa253/rpa_quaestvm-1.6.4.tar.gz
source
sdist
null
false
1a8acd966c53de83c80c2926202716e0
313ef5a595437aafd157031994ed18c14b862897e1b3a99fdd987e8c30771724
52b1a15617ec28a679a2929fd2c0efdea89a0db7bd74ceb9f0f4624eecbfa253
MIT
[ "LICENSE" ]
255
2.4
a2a-spec
0.1.0
The open specification for testing, validating, and guaranteeing agent-to-agent interactions.
<p align="center"> <strong>a2a-spec</strong><br> <em>The open specification for testing, validating, and guaranteeing agent-to-agent interactions.</em> </p> <p align="center"> <a href="https://pypi.org/project/a2a-spec/"><img src="https://img.shields.io/pypi/v/a2a-spec?style=flat-square" alt="PyPI"></a> <a href="https://github.com/padobrik/a2a-spec/actions/workflows/ci.yml"><img src="https://img.shields.io/github/actions/workflow/status/padobrik/a2a-spec/ci.yml?branch=main&style=flat-square&label=CI" alt="CI"></a> <a href="https://github.com/padobrik/a2a-spec/blob/main/LICENSE"><img src="https://img.shields.io/badge/license-Apache%202.0-blue?style=flat-square" alt="License"></a> <img src="https://img.shields.io/badge/python-3.11%20%7C%203.12%20%7C%203.13-blue?style=flat-square" alt="Python 3.11+"> <a href="https://github.com/padobrik/a2a-spec"><img src="https://img.shields.io/badge/typed-PEP%20561-brightgreen?style=flat-square" alt="Typed"></a> </p> --- ## The Problem Multi-agent AI systems are **impossible to test reliably**. When Agent A changes its output format, Agent B silently breaks. LLM outputs are non-deterministic, so CI pipelines either skip testing or flake constantly. Existing tools focus on prompt evaluation or observability — none provide **contract testing between agents**. ## The Solution **a2a-spec** is a specification, testing, and validation layer for multi-agent systems. Define what one agent expects from another as a YAML spec. Record LLM outputs as snapshots. Replay them deterministically in CI with zero LLM calls. Detect structural and semantic regressions before they reach production. ``` Agent A ──[spec]──> Agent B ──[spec]──> Agent C │ │ │ └── snapshot ──> replay ──> validate ──> ✓ CI passes ``` ## What a2a-spec is NOT | a2a-spec is **not** | Examples | What a2a-spec **is** | |---|---|---| | An agent framework | LangChain, CrewAI, AutoGen | A **testing layer** that sits alongside any framework | | An observability tool | LangSmith, Arize, Langfuse | A **validation engine** that runs in CI, not production | | A prompt evaluation tool | Promptfoo, DeepEval | A **contract testing** system between agents | | An agent runtime | n/a | A **specification framework** for agent boundaries | --- ## Quick Start ### Install ```bash pip install a2a-spec ``` With optional features: ```bash pip install a2a-spec[semantic] # Embedding-based semantic comparison pip install a2a-spec[langchain] # LangChain adapter pip install a2a-spec[dev] # Testing and linting tools pip install a2a-spec[all] # Everything ``` ### Initialize a project ```bash a2aspec init --name my-project ``` This creates: ``` my-project/ ├── a2a-spec.yaml # Project configuration └── a2a_spec/ ├── specs/ # Agent-to-agent contracts │ └── example-spec.yaml ├── snapshots/ # Recorded outputs (committed to git!) ├── scenarios/ # Test input scenarios └── adapters/ # Agent wrappers ``` ### Define a spec A spec is a YAML contract between a **producer** agent and a **consumer** agent. It defines structural, semantic, and policy requirements: ```yaml # a2a_spec/specs/triage-to-resolution.yaml spec: name: triage-to-resolution version: "1.0" producer: triage-agent consumer: resolution-agent description: "What the resolution agent expects from triage" structural: type: object required: [category, summary, confidence] properties: category: type: string enum: [billing, shipping, product, general] summary: type: string minLength: 10 maxLength: 500 confidence: type: number minimum: 0.0 maximum: 1.0 semantic: - rule: summary_reflects_input description: "Summary must faithfully reflect the customer message" method: embedding_similarity threshold: 0.8 policy: - rule: no_pii description: "Output must not contain PII" method: regex patterns: - '\b\d{4}[- ]?\d{4}[- ]?\d{4}[- ]?\d{4}\b' # Credit card - '\b\d{3}-\d{2}-\d{4}\b' # SSN ``` ### Record snapshots ```bash a2aspec record # Calls live agents via adapters, saves outputs to disk ``` Snapshots are JSON files committed to git — they become your deterministic test baselines. ### Test in CI (zero LLM calls) ```bash a2aspec test --replay # Validates saved snapshots against specs ``` No API keys needed. No LLM costs. Fully deterministic. Runs in milliseconds. ### Detect semantic drift After changing a prompt or upgrading a model: ```bash a2aspec record # Re-record with the new configuration a2aspec diff # Compare new vs. baseline outputs ``` The diff engine reports structural changes (fields added/removed/type-changed) and semantic drift (meaning shifted beyond threshold), with severity levels from LOW to CRITICAL. --- ## Core Concepts | Concept | Description | |---------|-------------| | **Spec** | A YAML file defining what one agent expects from another — structure, semantics, and policy rules | | **Snapshot** | A recorded LLM output for a given input, stored as JSON and committed to git | | **Replay** | Running validation against saved snapshots with zero LLM calls — fast, free, deterministic | | **Diff** | Structural + semantic comparison between old and new agent outputs, with severity levels | | **Pipeline** | A DAG of agents with routing conditions, tested end-to-end with spec validation at each step | | **Adapter** | A wrapper around your agent (function, HTTP, LangChain) so a2a-spec can call it | → See [docs/concepts.md](docs/concepts.md) for detailed explanations. --- ## Adapters — Wrap Any Agent a2a-spec is **framework-agnostic**. Adapters wrap your agents so the framework can call them during recording and testing. ### Plain async functions ```python from a2a_spec import FunctionAdapter async def my_triage_agent(input_data: dict) -> dict: # Your agent logic (calls OpenAI, Anthropic, local model, etc.) return {"category": "billing", "summary": "Customer reports duplicate charge", "confidence": 0.95} adapter = FunctionAdapter( fn=my_triage_agent, agent_id="triage-agent", version="1.0.0", model="gpt-4", ) ``` ### HTTP endpoints ```python from a2a_spec import HTTPAdapter adapter = HTTPAdapter( url="http://localhost:8000/triage", agent_id="triage-agent", version="1.0.0", headers={"Authorization": "Bearer $TOKEN"}, timeout=30.0, ) ``` ### Custom adapters ```python from a2a_spec import AgentAdapter, AgentMetadata, AgentResponse class MyCrewAIAdapter(AgentAdapter): def get_metadata(self) -> AgentMetadata: return AgentMetadata(agent_id="my-crew-agent", version="1.0") async def call(self, input_data: dict) -> AgentResponse: result = await my_crew.kickoff(input_data) return AgentResponse(output=result.dict()) ``` → See [docs/writing-adapters.md](docs/writing-adapters.md) for the full guide. --- ## Pipeline Testing Test entire multi-agent pipelines as a DAG. a2a-spec validates each agent's output against its spec and checks routing conditions: ```yaml pipeline: name: customer-support agents: triage-agent: {} billing-agent: {} shipping-agent: {} resolution-agent: {} edges: - from: triage-agent to: billing-agent condition: "output.category == 'billing'" - from: triage-agent to: shipping-agent condition: "output.category == 'shipping'" - from: [billing-agent, shipping-agent] to: resolution-agent test_cases: - name: billing_flow input: { message: "I was charged twice" } ``` ```bash a2aspec pipeline test pipeline.yaml --mode replay ``` → See [docs/architecture.md](docs/architecture.md) for the pipeline execution model. --- ## Configuration Project configuration lives in `a2a-spec.yaml`: ```yaml project_name: "my-project" version: "1.0" specs_dir: "./a2a_spec/specs" scenarios_dir: "./a2a_spec/scenarios" semantic: provider: sentence-transformers model: all-MiniLM-L6-v2 # Lazy-loaded, only when needed enabled: true storage: backend: local path: ./a2a_spec/snapshots ci: fail_on_semantic_drift: true drift_threshold: 0.15 replay_mode: exact ``` --- ## Python API Use a2a-spec programmatically in your existing test suite: ```python from a2a_spec import load_spec, validate_output, SnapshotStore, ReplayEngine # Load and validate spec = load_spec("a2a_spec/specs/triage-to-resolution.yaml") result = validate_output( {"category": "billing", "summary": "Customer charged twice", "confidence": 0.95}, spec, ) assert result.passed # Replay snapshots store = SnapshotStore("./a2a_spec/snapshots") engine = ReplayEngine(store) output = engine.replay("triage-agent", "billing_overcharge") # Diff two outputs from a2a_spec import DiffEngine diff = DiffEngine() results = diff.diff(old_output, new_output, semantic_threshold=0.85) for r in results: print(f"{r.field}: {r.severity} — {r.explanation}") # Policy enforcement from a2a_spec.policy.engine import PolicyEngine from a2a_spec.policy.builtin import no_pii_in_output engine = PolicyEngine() engine.register_validator("no_pii", no_pii_in_output) ``` --- ## CLI Reference | Command | Description | |---------|-------------| | `a2aspec init [DIR]` | Scaffold a new a2a-spec project with examples | | `a2aspec record` | Record live agent outputs as snapshots | | `a2aspec test --replay` | Validate snapshots against specs (deterministic, zero LLM calls) | | `a2aspec test --live` | Validate live agent outputs against specs | | `a2aspec diff` | Compare current outputs against baselines | | `a2aspec diff --agent NAME` | Diff a specific agent only | | `a2aspec pipeline test FILE` | Test a multi-agent pipeline DAG | | `a2aspec --version` | Show version | → See [docs/cli-reference.md](docs/cli-reference.md) for full options and flags. --- ## CI Integration a2a-spec is designed for CI-first workflows: ```yaml # .github/workflows/a2a-spec.yml name: Agent Contract Tests on: [push, pull_request] jobs: spec-tests: runs-on: ubuntu-latest steps: - uses: actions/checkout@v4 - uses: actions/setup-python@v5 with: python-version: "3.12" - run: pip install a2a-spec - run: a2aspec test --replay ``` **Key principle:** Record locally (with API keys), test in CI (with snapshots). Snapshots are committed to git — they are your test baselines. | Output Format | Flag | Use Case | |---|---|---| | Console (Rich) | `--format console` | Local development | | Markdown | `--format markdown` | PR comments | | JUnit XML | `--format junit` | CI test reporters | → See [docs/ci-integration.md](docs/ci-integration.md) for GitHub Actions, Jenkins, and more. --- ## Comparison | Feature | a2a-spec | Pact | DeepEval | Promptfoo | LangSmith | |---------|----------|------|----------|-----------|-----------| | Agent-to-agent contracts | ✅ | ✅ | ❌ | ❌ | ❌ | | LLM output snapshots | ✅ | ❌ | ❌ | ❌ | ❌ | | Deterministic CI replay | ✅ | ✅ | ❌ | ❌ | ❌ | | Semantic drift detection | ✅ | ❌ | ✅ | ✅ | ✅ | | Policy enforcement (PII, etc.) | ✅ | ❌ | ✅ | ✅ | ❌ | | Pipeline DAG testing | ✅ | ❌ | ❌ | ❌ | ❌ | | Framework agnostic | ✅ | ✅ | ❌ | ❌ | ❌ | | Zero LLM calls in CI | ✅ | N/A | ❌ | ❌ | ❌ | | Typed Python API (PEP 561) | ✅ | N/A | ✅ | N/A | ✅ | --- ## Architecture ``` src/a2a_spec/ ├── cli/ # Typer CLI (init, record, test, diff, pipeline) ├── spec/ # Spec schema (Pydantic), YAML loader, JSON Schema validator ├── snapshot/ # Record, store, fingerprint, and replay engine ├── diff/ # Structural (JSON) + semantic (embedding) comparison ├── pipeline/ # DAG builder, topological executor, execution traces ├── adapters/ # Agent wrappers: function, HTTP, LangChain ├── policy/ # Policy engine with regex and custom validators ├── semantic/ # Embedding model interface (sentence-transformers) ├── reporting/ # Console (Rich), Markdown, JUnit XML, GitHub annotations ├── config/ # YAML config loader with Pydantic validation ├── _internal/ # SHA256 hashing, safe expression evaluator, type aliases └── exceptions.py # Hierarchical error types with actionable messages ``` → See [docs/architecture.md](docs/architecture.md) for the full design. --- ## Examples The [`examples/customer_support/`](examples/customer_support/) directory contains a complete walkthrough: - Two agents (triage + resolution) with a2a-spec contract - YAML spec with structural, semantic, and policy rules - Pre-recorded snapshot for deterministic replay - Test scenarios and pytest integration - Step-by-step README --- ## Documentation | Guide | Description | |-------|-------------| | [Getting Started](docs/getting-started.md) | Installation and first test in 2 minutes | | [Core Concepts](docs/concepts.md) | Specs, snapshots, replay, diff explained | | [CLI Reference](docs/cli-reference.md) | Every command with all options | | [Writing Specs](docs/writing-specs.md) | Structural, semantic, and policy rules | | [Writing Adapters](docs/writing-adapters.md) | Wrap any agent for a2a-spec | | [CI Integration](docs/ci-integration.md) | GitHub Actions, JUnit, exit codes | | [Architecture](docs/architecture.md) | Module design and extension points | --- ## Contributing Contributions are welcome. See [CONTRIBUTING.md](CONTRIBUTING.md) for the development setup, check commands, and PR process. --- ## License Apache 2.0 — see [LICENSE](LICENSE) for details.
text/markdown
null
Fedor Kabachenko <fkabachenko@gmail.com>
null
null
Apache-2.0
a2a, agents, ai, contract-testing, llm, multi-agent, testing
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Sci...
[]
null
null
>=3.11
[]
[]
[]
[ "httpx<1.0,>=0.27", "jsonschema<5.0,>=4.20", "pydantic<3.0,>=2.0", "pyyaml<7.0,>=6.0", "rich<14.0,>=13.0", "typer<1.0,>=0.12", "langchain-core<1.0,>=0.2; extra == \"all\"", "mypy<2.0,>=1.10; extra == \"all\"", "numpy<3.0,>=1.26; extra == \"all\"", "pre-commit<4.0,>=3.7; extra == \"all\"", "pytes...
[]
[]
[]
[ "Homepage, https://github.com/padobrik/a2a-spec", "Documentation, https://github.com/padobrik/a2a-spec/tree/main/docs", "Repository, https://github.com/padobrik/a2a-spec", "Issues, https://github.com/padobrik/a2a-spec/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:25:32.631293
a2a_spec-0.1.0.tar.gz
61,555
48/ca/95b11d0442e24fb9fcc5883b8a67fd3a54c28642ca13e7474e5459bcad25/a2a_spec-0.1.0.tar.gz
source
sdist
null
false
4e535122e37cfde4bf42ee2ec67f6020
65747c3b3d66e4ee46e23519d8ec964fed79c9f60e1e669bc4cb3990364f9e16
48ca95b11d0442e24fb9fcc5883b8a67fd3a54c28642ca13e7474e5459bcad25
null
[ "LICENSE" ]
283
2.1
pcrm-book
1.0.7
Accompanying Python code to the Portfolio Construction and Risk Management book by Anton Vorobets.
[![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/fortitudo-tech/pcrm-book/HEAD?urlpath=%2Fdoc%2Ftree%2F%2Fcode) # Portfolio Construction and Risk Management book's Python code This repository contains the accompanying code to Portfolio Construction and Risk Management book © 2025 by Anton Vorobets. [You can find the latest PDF version of the book in this Substack post](https://antonvorobets.substack.com/p/pcrm-book). For a quick video introduction to what you can expect from this book and some fundamental perspectives, [watch this video](https://antonvorobets.substack.com/p/anton-vorobets-next-generation-investment-framework). Subscribe to the [Quantamental Investing Substack publication](https://antonvorobets.substack.com) to stay updated on all news related to the book. You can still support the project through [buy me a coffee](https://buymeacoffee.com/antonvorobets) or [Substack](https://antonvorobets.substack.com). # Applied Quantitative Investment Management course You can access [a course that carefully goes through the book and its accompanying code](https://antonvorobets.substack.com/t/course) from this repository. [Read more about the course and how you get access here](https://antonvorobets.substack.com/p/course-q-and-a). # Running the code It is recommended to install the book's code dependencies in a [conda environment](https://conda.io/projects/conda/en/latest/user-guide/concepts/environments.html). After cloning the repository to your local machine, you can install the dependencies using the following command in your terminal: conda env create -f environment.yml You can then activate the conda environment and start a [JupyterLab](https://jupyter.org/) instance using the following commands: conda activate pcrm-book jupyter lab If you are completely new to [conda environments](https://conda.io/projects/conda/en/latest/user-guide/tasks/manage-environments.html) and Jupyter notebooks, you can find a lot of information online. You can also run the code without any local installations using [Binder](https://mybinder.org/v2/gh/fortitudo-tech/pcrm-book/HEAD?urlpath=%2Fdoc%2Ftree%2F%2Fcode). Note however that Binder servers are not always available and might have insufficient resources to run all the examples. # Feedback Please post your feedback in the community Discussions forum. I will try to incorporate the feedback in the book. See the book's preface for some general perspectives on what it tries to achieve, and which kind feedback will be considered appropriate. # Thank you for your support Your support made it possible for this book to be written. Besides your personal monetary support, you can help improve the quality of the book by simply publicly sharing your positive experience with the book and its code, thereby encouraging more people to support the project. You are also encouraged to give this and the supporting [fortitudo.tech](https://github.com/fortitudo-tech/fortitudo.tech) repository a star. No matter how much economic support this project realistically gets, it will only be a small fraction of the opportunity costs from writing the book and making it freely available online. Hence, you are encouraged to support it by the amount that you think it is worth to you. [If you claim one of the significiant contributor perks](https://igg.me/at/pcrm-book), you can choose to be recognized in the book's preface. You will additionally get a one-year paid complimentary Substack subscription to the [Quantamental Investing publication](https://antonvorobets.substack.com), which contains exclusive case studies and allow you to continue asking questions. # Licenses The Portfolio Construction and Risk Management book © 2025 by Anton Vorobets is licensed under Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International. To view a copy of this license, visit https://creativecommons.org/licenses/by-nc-nd/4.0/ The accompanying code to the Portfolio Construction and Risk Management book © 2025 by Anton Vorobets is licensed under version 3 of the GNU General Public License. To view a copy of this license, visit https://www.gnu.org/licenses/gpl-3.0.en.html
text/markdown
Anton Vorobets
admin@fortitudo.tech
null
null
GPL-3.0-or-later
CVaR, Entropy Pooling, Quantitative Finance, Portfolio Optimization, Risk Management
[ "Intended Audience :: Education", "Intended Audience :: Financial and Insurance Industry", "Intended Audience :: Science/Research", "License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.9", "Programming L...
[]
https://pcrmbook.com
null
<3.14,>=3.9
[]
[]
[]
[ "fortitudo-tech<2.0.0,>=1.1.8", "jupyterlab<5.0.0,>=4.3.0", "yfinance<0.3.0,>=0.2.50", "seaborn<0.14.0,>=0.13.0" ]
[]
[]
[]
[ "Repository, https://github.com/fortitudo-tech/pcrm-book", "Documentation, https://antonvorobets.substack.com/p/pcrm-book", "Issues, https://github.com/fortitudo-tech/pcrm-book/issues" ]
poetry/1.3.2 CPython/3.8.10 Linux/5.15.167.4-microsoft-standard-WSL2
2026-02-19T12:25:28.767362
pcrm_book-1.0.7.tar.gz
15,787
d1/07/8e0c06702facf0a527cdfa87be84d08832e685bf12a1fd8224ec40f26daa/pcrm_book-1.0.7.tar.gz
source
sdist
null
false
2a99df7845fcc909bc509da2a641c05e
e0dd163e4b0684a1816d70c08ee46693526fadf22db10ee5412b9f4bea19f9c9
d1078e0c06702facf0a527cdfa87be84d08832e685bf12a1fd8224ec40f26daa
null
[]
255
2.4
mathpacki
0.3.0
A simple Python package for numeric, set, and equation operations (number; set; equation: solve linear, add/subtract/scale equations)
# mathpacki A lightweight Python package for numeric and set operations. ## Installation ```bash pip install mathpacki ``` ## Usage ### Modular imports ```python from mathpacki import numbers from mathpacki import sets from mathpacki import equations ``` ### Number operations ```python from mathpacki import numbers numbers.add(2, 3) # 5 numbers.subtract(10, 4) # 6 numbers.multiply(3, 7) # 21 numbers.divide(15, 3) # 5.0 numbers.power(2, 8) # 256 numbers.modulo(17, 5) # 2 ``` Or import individual functions: ```python from mathpacki.numbers import add, multiply add(2, 3) # 5 multiply(3, 7) # 21 ``` ### Set operations ```python from mathpacki import sets a, b = {1, 2, 3}, {2, 3, 4} sets.union(a, b) # {1, 2, 3, 4} sets.intersection(a, b) # {2, 3} sets.difference(a, b) # {1} sets.symmetric_difference(a, b) # {1, 4} sets.is_subset({1, 2}, a) # True sets.is_superset(a, {1, 2}) # True sets.is_disjoint({1, 2}, {3, 4}) # True ``` Or import individual functions: ```python from mathpacki.sets import union, intersection union({1, 2}, {2, 3}) # {1, 2, 3} ``` ### Equation operations (linear: a*x + b = c) Equations are represented as tuples `(a, b, c)` for **a·x + b = c**. ```python from mathpacki import equations # Solve a*x + b = c for x equations.solve_linear(2, -3, 7) # 2x - 3 = 7 -> x = 5.0 equations.solve_linear(1, 0, 4) # x = 4 -> 4.0 # Evaluate left-hand side at x equations.evaluate_linear(2, -3, 5) # 2*5 - 3 = 7.0 # Operate on equations (each eq = (a, b, c) for a*x + b = c) eq1 = (2, 1, 5) # 2x + 1 = 5 eq2 = (1, -1, 1) # x - 1 = 1 equations.add_equations(eq1, eq2) # (3, 0, 6) -> 3x = 6 equations.subtract_equations(eq1, eq2) # (1, 2, 4) -> x + 2 = 4 equations.scale_equation(eq1, 2) # (4, 2, 10) -> 4x + 2 = 10 ``` ## Requirements - Python 3.8+ ## Publishing to PyPI 1. **Create a PyPI account** at [pypi.org](https://pypi.org/account/register/). 2. **Install build tools**: ```bash pip install --upgrade build twine ``` 3. **Update `pyproject.toml`**: Set your name, email, and project URLs (and change the package `name` if `mathpack` is already taken on PyPI). 4. **Build the package**: ```bash python -m build ``` This creates `dist/` with a `.whl` and `.tar.gz` file. 5. **Upload to PyPI** (you will be prompted for your PyPI username and password or API token): ```bash python -m twine upload dist/* ``` For first-time testing, use TestPyPI: `python -m twine upload --repository testpypi dist/*` ## License MIT
text/markdown
null
Samyar Modabber <samyar.modabber@gmail.com>
null
null
null
arithmetic, calculator, math, operations
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Pyt...
[]
null
null
>=3.8
[]
[]
[]
[]
[]
[]
[]
[ "Homepage, https://github.com/samyarmodabber/mathpack", "Repository, https://github.com/samyarmodabber/mathpack" ]
twine/6.2.0 CPython/3.11.9
2026-02-19T12:24:25.482040
mathpacki-0.3.0.tar.gz
4,074
5a/8b/b6f6b4658853d2012f5a8dc9ffaea0aea33d7bc90ab59dbf964780fab78a/mathpacki-0.3.0.tar.gz
source
sdist
null
false
da08802b875de6708487418eb0962c2d
7eb2b91a2ece8f8a512f68aa652e775510f2c707c6ac43363b00c34038e910ac
5a8bb6f6b4658853d2012f5a8dc9ffaea0aea33d7bc90ab59dbf964780fab78a
MIT
[ "LICENSE" ]
284
2.3
SQLModel-translation
0.1.0
Translation library for SQLModel and FastAPI
# SQLModel-translation SQLModel-translation is a translation library for [SQLModel](https://sqlmodel.tiangolo.com) and [FastAPI](https://fastapi.tiangolo.com). This project uses [uv](https://docs.astral.sh/uv/) for package managment. To generate the documentation run `make docs` and visit http://127.0.0.1:8000/. For more actions see the the Makefile in this directory. Running `make` will print out all the targets with descriptions.
text/markdown
null
null
null
null
null
null
[]
[]
null
null
>=3.12
[]
[]
[]
[ "fastapi>=0.119.1", "sqlalchemy>=2.0.44", "sqlmodel>=0.0.27" ]
[]
[]
[]
[]
uv/0.10.4 {"installer":{"name":"uv","version":"0.10.4","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Arch Linux","version":null,"id":null,"libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}
2026-02-19T12:24:00.604737
sqlmodel_translation-0.1.0.tar.gz
5,583
8e/7e/b6e0035d7e8c7ae29070cf496c0be2d90e8626c38df4434042b60e4cd9f3/sqlmodel_translation-0.1.0.tar.gz
source
sdist
null
false
a942ef3889a89fc5335fd4d2f286da6e
ed89e1231c8b73747de72757151b17fbe915a89288fb4db2679095501dc809e3
8e7eb6e0035d7e8c7ae29070cf496c0be2d90e8626c38df4434042b60e4cd9f3
null
[]
0
2.1
sbcommons
1.2.3
Packages shared between several Data related systems in Haypp Group
# sbcommons [![CI](https://github.com/Snusbolaget/sbcommons/actions/workflows/ci.yml/badge.svg)](https://github.com/Snusbolaget/sbcommons/actions/workflows/ci.yml) [![PyPI version](https://badge.fury.io/py/sbcommons.svg)](https://pypi.org/project/sbcommons/) A comprehensive Python library providing shared utilities and clients for AWS services, CRM integrations (Klaviyo, Symplify), messaging platforms (Slack, Microsoft Teams), and data processing operations used across Haypp Group's data systems. ## Features - **AWS Integration**: Clients for S3, Redshift, Lambda, DMS, SNS, and Secrets Manager - **CRM Clients**: - Klaviyo API v3 client with comprehensive campaign, event, and customer management - Symplify integration for customer relationship management - ADP client for survey data - **Messaging**: Slack and Microsoft Teams webhook utilities with modern Adaptive Cards support - **Logging**: Enhanced logging with Lambda-compatible logger and rotating file handlers - **Data Utilities**: Extract and parse utilities for common data transformation tasks ## Installation Install from PyPI: ```bash pip install sbcommons ``` ## Requirements - Python 3.9, 3.10, 3.11, or 3.12 - boto3 >= 1.24.35 - requests >= 2.28.1 - Other dependencies listed in `requirements.txt` ## Usage ### AWS Services ```python from sbcommons.aws.s3 import S3Client from sbcommons.aws.redshift import RedshiftClient from sbcommons.aws.secrets import get_secret # S3 operations s3_client = S3Client() s3_client.upload_file('local.txt', 'bucket', 'key.txt') # Redshift queries redshift = RedshiftClient(host='your-host', database='db') results = redshift.execute_query('SELECT * FROM table') # Secrets Manager secret = get_secret('my-secret-name') ``` ### CRM Integration ```python from sbcommons.crm.klaviyo.client import KlaviyoClient # Klaviyo operations klaviyo = KlaviyoClient(api_key='your-key') campaigns = klaviyo.get_email_campaigns() events = klaviyo.get_events(metric_id='metric123') ``` ### Messaging ```python from sbcommons.teams.teams import post_to_teams from sbcommons.slack.slack import SlackMessenger # Microsoft Teams (Adaptive Cards) post_to_teams('https://webhook-url', 'Message title', 'Message body') # Slack slack = SlackMessenger('https://slack-webhook-url') slack.post_message('Hello from sbcommons!') ``` ## Development ### Setup Development Environment Clone the repository and create a virtual environment: ```bash git clone https://github.com/Snusbolaget/sbcommons.git cd sbcommons pip install -r requirements.txt ``` ### Running Tests ```bash # Run tests with pytest pytest tests/ -v --cov=sbcommons --cov-report=term ``` Tests are automatically run via GitHub Actions on push and pull requests for Python 3.9, 3.10, and 3.11. ### Building and Publishing #### Local Development Builds For local testing, you can build the package: ```bash # Install build tools pip install build twine # Build distributions python -m build # Validate the build twine check dist/* # Clean build artifacts when done rm -rf dist/ build/ *.egg-info ``` #### Testing on TestPyPI (Optional) Before releasing to production, you can test the package on TestPyPI: ```bash # Build and upload to TestPyPI python -m build twine upload --repository testpypi dist/* # Test installation from TestPyPI pip install --index-url https://test.pypi.org/simple/ sbcommons ``` **Note**: TestPyPI uploads are for manual testing only and should not be part of the automated workflow. #### Publishing to PyPI (Production) **Publishing is fully automated via GitHub Actions.** When your PR is merged to `main`, the release workflow automatically: 1. Builds the package distributions 2. Publishes to PyPI using the `PYPI_API_TOKEN` secret **To release a new version:** 1. Update the version in `setup.py` 2. Document changes in `CHANGELOG.md` 3. Create a PR to `main` 4. Once merged, the package is automatically published to PyPI **Manual publishing is not recommended** to ensure consistency and prevent version conflicts. All releases should go through the PR review process. ## Project Structure ``` sbcommons/ ├── adp/ # ADP client integration ├── aws/ # AWS service clients (S3, Redshift, Lambda, etc.) ├── crm/ # CRM integrations (Klaviyo, Symplify) ├── extract_utils/ # Data extraction utilities ├── logging/ # Enhanced logging utilities ├── messaging/ # Webhook utilities ├── parse_utils/ # Configuration and text parsing ├── slack/ # Slack messaging client └── teams/ # Microsoft Teams messaging (Adaptive Cards) ``` ## CI/CD The project uses GitHub Actions for continuous integration and deployment: - **CI Workflow** (`.github/workflows/ci.yml`): Runs tests on Python 3.9-3.11 for all pull requests and pushes to main - **Release Workflow** (`.github/workflows/release.yml`): Automatically publishes to PyPI when changes are pushed to main branch ## Contributing 1. Create a feature branch from `main` 2. Make your changes and add tests 3. Update the CHANGELOG.md with your changes 4. Update the version in `setup.py` 5. Submit a pull request The CI workflow will verify that: - Tests pass on all supported Python versions - If `sbcommons/` code is modified, both CHANGELOG.md and setup.py versions are updated ## Changelog See [CHANGELOG.md](CHANGELOG.md) for detailed version history. **Maintained by Haypp Group Data Team** | [data@hayppgroup.com](mailto:data@hayppgroup.com)
text/markdown
Haypp Group
data@hayppgroup.com
null
null
null
null
[ "Programming Language :: Python :: 3.9", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent" ]
[]
https://github.com/Snusbolaget
null
null
[]
[]
[]
[]
[]
[]
[]
[]
twine/6.2.0 CPython/3.9.25
2026-02-19T12:23:18.184755
sbcommons-1.2.3.tar.gz
44,524
21/40/5f68715469179581b9c879f6530a84decdfc854d3833201f4e1c144272e7/sbcommons-1.2.3.tar.gz
source
sdist
null
false
dd25d6a903a22d29ae331cdd5a756213
cbe78cd4534b1d31517a5369ac58035700fc7a320845766ba01233520b47f5fa
21405f68715469179581b9c879f6530a84decdfc854d3833201f4e1c144272e7
null
[]
279
2.4
9zx
0.1.1
Simple HiHello responder library
# HiHello HiHello
text/markdown
null
Oywe <ncql@proton.me>
null
null
MIT
null
[]
[]
null
null
>=3.8
[]
[]
[]
[]
[]
[]
[]
[]
twine/6.2.0 CPython/3.11.14
2026-02-19T12:23:16.539863
9zx-0.1.1.tar.gz
1,440
1a/5e/9c823c8ad1a256ac48ac3fca927ee67405ae3886c965ab59932b6f3abcb5/9zx-0.1.1.tar.gz
source
sdist
null
false
f6e0c100b0d17e08416ef9487d839134
841065bc1b833cf148d499f06135fce98cfc2d98bf1212f8df68c03bbdfcaec6
1a5e9c823c8ad1a256ac48ac3fca927ee67405ae3886c965ab59932b6f3abcb5
null
[]
307
2.4
sphinx-filter-tabs
1.3.1
A Sphinx extension for accessible, CSS-first filterable content tabs.
# Sphinx Filter Tabs Extension [![Tests and Docs Deployment](https://github.com/aputtu/sphinx-filter-tabs/actions/workflows/test.yml/badge.svg)](https://github.com/aputtu/sphinx-filter-tabs/actions/workflows/test.yml) [![PyPI version](https://img.shields.io/pypi/v/sphinx-filter-tabs.svg)](https://pypi.org/project/sphinx-filter-tabs/) [![PyPI - Python Version](https://img.shields.io/pypi/pyversions/sphinx-filter-tabs.svg)](https://pypi.org/project/sphinx-filter-tabs/) [![PyPI - License](https://img.shields.io/pypi/l/sphinx-filter-tabs.svg)](https://github.com/aputtu/sphinx-filter-tabs/blob/main/LICENSE) A robust Sphinx extension for creating accessible, filterable content tabs using pure CSS and semantic HTML. **📖 View extension and documentation at: https://aputtu.github.io/sphinx-filter-tabs/** This extension provides `filter-tabs` and `tab` directives to create user-friendly, switchable content blocks. Perfect for showing code examples in multiple languages, installation instructions for different platforms, or any content that benefits from organized, filterable presentation. ## Key Features - **Pure CSS Implementation:** Zero JavaScript dependencies for maximum compatibility and performance - **Fully Accessible:** WAI-ARIA compliant with native keyboard navigation and screen reader support - **Semantic HTML:** Uses standard form controls (radio buttons) for robust, predictable behavior - **Universal Compatibility:** Works in all environments where CSS is supported, including strict CSP policies - **Easy Customization:** Theme colors and styling through simple CSS custom properties - **Multiple Output Formats:** Graceful fallback to sequential content in PDF/LaTeX builds - **Proven Reliability:** Comprehensive test suite across multiple Python and Sphinx versions ## Quick Start ### Installation ```bash pip install sphinx-filter-tabs ``` ### Enable the Extension Add to your `conf.py`: ```python extensions = [ # ... your other extensions ... 'filter_tabs.extension', ] ``` ### Basic Usage ```rst .. filter-tabs:: This content appears above all tabs. .. tab:: Python Install using pip: .. code-block:: bash pip install my-package .. tab:: Conda (default) Install using conda: .. code-block:: bash conda install my-package .. tab:: From Source Build from source: .. code-block:: bash git clone https://github.com/user/repo.git cd repo pip install -e . ``` ## Configuration Options Add these optional settings to your `conf.py`: ```python # Customize the active tab highlight color filter_tabs_highlight_color = '#007bff' # Default: '#007bff' # Enable debug logging during development filter_tabs_debug_mode = False # Default: False ``` ## Advanced Usage ### Custom Legend Override the auto-generated legend: ```rst .. filter-tabs:: :legend: Select Your Installation Method .. tab:: Quick Install Content here... ``` ### ARIA Labels for Accessibility Provide descriptive labels for screen readers: ```rst .. filter-tabs:: .. tab:: CLI :aria-label: Command line installation instructions Content for command line users... ``` ### Nested Tabs Create complex layouts with nested tab groups: ```rst .. filter-tabs:: .. tab:: Windows Choose your package manager: .. filter-tabs:: .. tab:: Chocolatey choco install my-package .. tab:: Scoop scoop install my-package ``` ## How It Works This extension uses a **pure CSS architecture** with semantic HTML: - **Radio buttons** provide the selection mechanism (hidden but accessible) - **CSS `:checked` selectors** control panel visibility - **Fieldset/legend** structure provides semantic grouping - **ARIA attributes** enhance screen reader support - **Native keyboard navigation** works through standard form controls This approach ensures: - **Maximum compatibility** across all browsers and assistive technologies - **Better performance** with no JavaScript parsing or execution - **Enhanced security** for environments with strict Content Security Policies - **Simplified maintenance** with fewer dependencies and potential conflicts ## Browser Support Works in all modern browsers that support: - CSS3 selectors (`:checked`, attribute selectors) - Basic ARIA attributes - HTML5 form elements This includes all browsers from the last 10+ years. ## Development ### Quick Setup ```bash git clone https://github.com/aputtu/sphinx-filter-tabs.git cd sphinx-filter-tabs ./scripts/setup_dev.sh ``` This creates a virtual environment and builds the documentation. ### Development Commands ```bash # Activate virtual environment source venv/bin/activate # Run tests pytest # Build documentation ./scripts/dev.sh html # Run tests across multiple Sphinx versions tox # Clean build and start fresh ./scripts/dev.sh clean-all # Export project structure for analysis ./scripts/export-project.sh ``` ### Testing The project includes comprehensive tests covering: - Basic tab functionality and content visibility - Accessibility features and ARIA compliance - Nested tabs and complex layouts - Multiple output formats (HTML, LaTeX) - Cross-browser compatibility Tests run automatically on: - Python versions 3.10, 3.12 - Sphinx versions 7.0, 7.4, 8.0, 8.2, 9.0, 9.1 - Multiple operating systems via GitHub Actions ## Architecture The extension consists of three main components: - **`extension.py`** - Sphinx integration, directives, and node definitions - **`renderer.py`** - HTML generation and output formatting - **`static/filter_tabs.css`** - Pure CSS styling and functionality This clean separation makes the code easy to understand, test, and maintain. ## Contributing Contributions are welcome! Please: 1. Fork the repository 2. Create a feature branch 3. Add tests for new functionality 4. Ensure all tests pass: `pytest` 5. Submit a pull request ## License GNU General Public License v3.0. See [LICENSE](LICENSE) for details. ## Changelog See [CHANGELOG.md](docs/changelog.rst) for version history and migration notes. ## Support - **Documentation**: https://aputtu.github.io/sphinx-filter-tabs/ - **Issues**: https://github.com/aputtu/sphinx-filter-tabs/issues - **PyPI**: https://pypi.org/project/sphinx-filter-tabs/
text/markdown
null
Aputsiak Niels Janussen <aputtu+sphinx@gmail.com>
null
null
GNU General Public License v3.0
sphinx, extension, tabs, filter, documentation, css-only, accessibility, keyboard-navigation
[ "Development Status :: 5 - Production/Stable", "Framework :: Sphinx :: Extension", "Intended Audience :: Developers", "License :: OSI Approved :: GNU General Public License v3 (GPLv3)", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10",...
[]
null
null
>=3.10
[]
[]
[]
[ "Sphinx<10.0,>=7.0" ]
[]
[]
[]
[ "Homepage, https://github.com/aputtu/sphinx-filter-tabs", "Repository, https://github.com/aputtu/sphinx-filter-tabs.git", "Issues, https://github.com/aputtu/sphinx-filter-tabs/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:23:10.139376
sphinx_filter_tabs-1.3.1.tar.gz
33,843
2b/43/136c3a7878363f37eff04ff6af5d3f5180555f0dab4e3136f8a0d1f8ccca/sphinx_filter_tabs-1.3.1.tar.gz
source
sdist
null
false
7a8e4e42e0e18a79884976d00d9fe369
4f57e52fb739cd2880ab5faeb1be477627243b811bcb2df18b38b01954c2e1bd
2b43136c3a7878363f37eff04ff6af5d3f5180555f0dab4e3136f8a0d1f8ccca
null
[ "LICENSE" ]
294
2.4
songstats-sdk
0.1.0
Official Python client for the Songstats Enterprise API
# Songstats Python SDK Official Python client for the **Songstats Enterprise API**. 📚 API Documentation: https://docs.songstats.com 🔑 API Key Access: Please contact api@songstats.com --- ## Requirements - Python >= 3.10 --- ## Installation Install from PyPI: pip install songstats-sdk For local development: pip install -e ".[dev]" --- ## Quick Start ```python from songstats_sdk import SongstatsClient client = SongstatsClient(api_key="YOUR_API_KEY") # API status status = client.info.status() # Track information track = client.tracks.info(songstats_track_id="abcd1234") # Artist statistics artist_stats = client.artists.stats( songstats_artist_id="abcd1234", source="spotify", ) ``` --- ## Authentication All requests include your API key in the `apikey` header. You can generate an API key in your Songstats Enterprise dashboard. We recommend storing your key securely in environment variables: export SONGSTATS_API_KEY=your_key_here --- ## Available Resource Clients - `client.info` - `client.tracks` - `client.artists` - `client.collaborators` - `client.labels` Info endpoints: - `client.info.sources()` -> `/sources` - `client.info.status()` -> `/status` - `client.info.definitions()` -> `/definitions` --- ## Error Handling ```python from songstats_sdk import SongstatsAPIError, SongstatsTransportError try: client.tracks.info(songstats_track_id="invalid") except SongstatsAPIError as exc: print(f"API error: {exc}") except SongstatsTransportError as exc: print(f"Transport error: {exc}") ``` --- ## Development To work on the SDK locally: git clone https://github.com/songstats/songstats-python-sdk.git cd songstats-python-sdk pip install -e ".[dev]" pytest --- ## Versioning This SDK follows Semantic Versioning (SemVer). --- ## License MIT
text/markdown
Songstats
null
null
null
MIT
api, music, sdk, songstats
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Topic :: Software Develo...
[]
null
null
>=3.10
[]
[]
[]
[ "httpx<1.0.0,>=0.27.0", "pytest<9.0.0,>=8.0.0; extra == \"dev\"", "respx<1.0.0,>=0.21.0; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://songstats.com", "Documentation, https://docs.songstats.com" ]
twine/6.2.0 CPython/3.12.4
2026-02-19T12:23:06.284424
songstats_sdk-0.1.0.tar.gz
7,911
68/7f/d44468d8b3174ae15bd0e976d6f0728cad1daac1e93ce5557b9c39e1c10e/songstats_sdk-0.1.0.tar.gz
source
sdist
null
false
a6875fe296cfcd15d4641ff80d3d4190
c105d2cd82a92e32d9b1f075c359ec7799e8c26f27e099c5c537b67a8d39fe85
687fd44468d8b3174ae15bd0e976d6f0728cad1daac1e93ce5557b9c39e1c10e
null
[ "LICENSE" ]
264
2.4
aigie
0.2.14
Enterprise-grade AI agent reliability monitoring and autonomous remediation
# Aigie SDK Production-grade Python SDK for integrating Aigie monitoring into your AI agent workflows. ## ✨ Features - 🚀 **Event Buffering**: 10-100x performance improvement with batch uploads - 🎯 **Decorator Support**: 50%+ less boilerplate code - ⚙️ **Flexible Configuration**: Config class with sensible defaults - 🔄 **Automatic Retries**: Exponential backoff with configurable policies - 🔗 **LangChain Integration**: Seamless callback handler - 📊 **Production Ready**: Handles network failures, race conditions, and more ## Quick Start ### Installation ```bash pip install aigie ``` ### Basic Usage #### Option 1: Context Manager (Traditional) ```python from aigie import Aigie aigie = Aigie() await aigie.initialize() async with aigie.trace("My Workflow") as trace: async with trace.span("operation", type="llm") as span: result = await do_work() span.set_output({"result": result}) ``` #### Option 2: Decorator (Recommended - 50% less code!) ```python from aigie import Aigie aigie = Aigie() await aigie.initialize() @aigie.trace(name="my_workflow") async def my_workflow(): @aigie.span(name="operation", type="llm") async def operation(): return await do_work() return await operation() ``` #### Option 3: With Configuration ```python from aigie import Aigie, Config config = Config( aigie_url="https://portal.aigie.io/api", aigie_token="your-token", # Required for data to be sent batch_size=100, # Buffer 100 events before sending flush_interval=5.0 # Or flush every 5 seconds ) aigie = Aigie(config=config) await aigie.initialize() ``` ## Configuration ### Environment Variables ```bash export AIGIE_TOKEN=your-token-here # Required for data to be sent export AIGIE_URL=https://portal.aigie.io/api export AIGIE_BATCH_SIZE=100 export AIGIE_FLUSH_INTERVAL=5.0 ``` ### Config Object ```python from aigie import Config config = Config( aigie_url="https://portal.aigie.io/api", aigie_token="your-token", # Required for data to be sent batch_size=100, flush_interval=5.0, enable_buffering=True, # Default: True max_retries=3 ) ``` ### Module-level Configuration (LiteLLM-style) ```python import aigie aigie.aigie_token = "your-token" # Required for data to be sent aigie.aigie_url = "https://portal.aigie.io/api" aigie.init() # Initialize with module-level settings ``` ## Performance ### Before (No Buffering) - 1000 spans = 1000+ API calls - ~30 seconds total time - High network overhead ### After (With Buffering) - 1000 spans = 2-10 API calls - ~0.5 seconds total time - **99%+ reduction in API calls** ## Advanced Features ### OpenTelemetry Integration Works with any OpenTelemetry-compatible tool (Datadog, New Relic, Jaeger, etc.): ```python from aigie import Aigie from aigie.opentelemetry import setup_opentelemetry aigie = Aigie() await aigie.initialize() # One-line setup setup_opentelemetry(aigie, service_name="my-service") # Now all OTel spans automatically go to Aigie! from opentelemetry import trace tracer = trace.get_tracer(__name__) with tracer.start_as_current_span("operation"): # Automatically traced pass ``` ### Synchronous API For non-async codebases: ```python from aigie import AigieSync aigie = AigieSync() aigie.initialize() # Blocking with aigie.trace("workflow") as trace: with trace.span("operation") as span: result = do_work() # Sync code span.set_output({"result": result}) ``` ## Installation ### Basic ```bash pip install aigie ``` ### With OpenTelemetry ```bash pip install aigie[opentelemetry] ``` ### With LangChain ```bash pip install aigie[langchain] ``` ### All Features ```bash pip install aigie[all] ``` ## Advanced Features (Phase 3) ### W3C Trace Context Propagation Distributed tracing across microservices: ```python # Extract from incoming request context = aigie.extract_trace_context(request.headers) async with aigie.trace("workflow") as trace: trace.set_trace_context(context) # Propagate to downstream service headers = trace.get_trace_headers() response = await httpx.get("https://api.example.com", headers=headers) ``` ### Prompt Management Create, version, and track prompts: ```python # Create prompt prompt = await aigie.prompts.create( name="customer_support", template="You are a helpful assistant. Customer: {customer_name}", version="1.0" ) # Use in trace async with aigie.trace("support") as trace: trace.set_prompt(prompt) rendered = prompt.render(customer_name="John") response = await llm.ainvoke(rendered) ``` ### Evaluation Hooks Automatic quality monitoring: ```python from aigie import EvaluationHook, ScoreType hook = EvaluationHook( name="accuracy", evaluator=accuracy_evaluator, score_type=ScoreType.ACCURACY ) async with aigie.trace("workflow") as trace: trace.add_evaluation_hook(hook) result = await do_work() await trace.run_evaluations(expected, result) ``` ### Streaming Support Real-time span updates: ```python async with aigie.trace("workflow") as trace: async with trace.span("llm_call", stream=True) as span: async for chunk in llm.astream("Hello"): span.append_output(chunk) # Update in real-time yield chunk ``` ## Documentation - [SDK Improvement Analysis](./SDK_IMPROVEMENT_ANALYSIS.md) - Comprehensive analysis - [Examples](./EXAMPLES_IMPROVED.md) - Before/after code examples - [Comparison Table](./COMPARISON_TABLE.md) - Feature comparison with market leaders - [Phase 2 Features](./PHASE2_FEATURES.md) - OpenTelemetry, Sync API, Type Hints - [Phase 3 Features](./PHASE3_FEATURES.md) - W3C Context, Prompts, Evaluations, Streaming
text/markdown
Aigie Team
support@aigie.io
null
null
null
ai agent monitoring observability llm reliability remediation
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Py...
[]
https://github.com/aigie/aigie-sdk
null
>=3.9
[]
[]
[]
[ "httpx>=0.25.0", "zstandard>=0.22.0; extra == \"compression\"", "openai>=1.0.0; extra == \"openai\"", "anthropic>=0.18.0; extra == \"anthropic\"", "google-generativeai>=0.3.0; extra == \"gemini\"", "langchain-core>=0.1.0; extra == \"langchain\"", "langgraph>=0.0.20; extra == \"langgraph\"", "langchain...
[]
[]
[]
[ "Documentation, https://docs.aigie.io", "Source, https://github.com/aigie/aigie-sdk", "Tracker, https://github.com/aigie/aigie-sdk/issues", "Changelog, https://github.com/aigie/aigie-sdk/blob/main/CHANGELOG.md" ]
twine/6.2.0 CPython/3.14.3
2026-02-19T12:22:42.083044
aigie-0.2.14.tar.gz
883,876
05/65/7a5eecc4ee6343f96074dc25ae7f3c6865b6f0e2eea347adc79848011a05/aigie-0.2.14.tar.gz
source
sdist
null
false
bfd4861a666495c41df376cc755d5c54
cbcb41fa94de5ee3f03db95467942dd4eb9fe9c53cb56ecfd7234ed202518e56
05657a5eecc4ee6343f96074dc25ae7f3c6865b6f0e2eea347adc79848011a05
null
[]
1,240
2.4
sum-cli
3.3.2
SUM Platform CLI: single control plane for site lifecycle management
# SUM CLI (v3.3.2) [![PyPI](https://img.shields.io/pypi/v/sum-cli.svg)](https://pypi.org/project/sum-cli/) The SUM CLI is the single control plane for deploying and managing SUM Platform client sites on staging and production servers. ## Install ```bash pip install sum-cli sum-platform --version ``` ### With Gitea Support If using Gitea instead of GitHub for repository hosting: ```bash pip install sum-cli[gitea] ``` ## Initial Setup Before using the CLI, configure your infrastructure settings: ```bash sudo sum-platform setup ``` This interactive command creates `/etc/sum/config.yml` with your staging/production server settings. ## Commands | Command | Description | Requires Sudo | |---------|-------------|---------------| | `setup` | Configure infrastructure settings | Yes | | `init` | Create new site at `/srv/sum/<name>/` | Yes | | `update` | Pull updates, migrate, restart | No (staging) | | `backup` | PostgreSQL backup via pgBackRest | No | | `restore` | Point-in-time database restore | No | | `destroy` | Tear down a site and all its infrastructure | Yes | | `monitor` | Check backup health, send alerts | No | | `promote` | Deploy staging site to production | No | | `check` | Validate project setup | No | | `theme` | Theme management (list, check, update) | No | | `themes` | List themes (deprecated, use `theme list`) | No | ## Pre-flight Checks Every state-mutating command (`init`, `promote`, `backup`, `restore`, `update`, `destroy`) runs pre-flight checks before executing. These read-only checks validate that prerequisites are met — missing config files, unreachable servers, and missing tokens are caught before any work begins. ### Example Output ``` Pre-flight checks: ──────────────────────────────────────────────────────────── ✅ Site directory: /srv/sum/acme exists ✅ PostgreSQL cluster: Cluster 'acme' is online ✅ SSH connectivity: SSH to 10.0.0.1:22 succeeded ✅ Disk space: 15.2 GB free on /srv/sum ❌ Gitea token: GITEA_TOKEN not set in environment ↳ Fix: Export GITEA_TOKEN or add it to /etc/sum/env ──────────────────────────────────────────────────────────── ``` If any required check fails, the command aborts with actionable fix hints. Recommended checks (e.g., `destroy`'s site existence check) warn but don't block. ### Skipping Pre-flight For automation or CI, bypass checks with the global flag or environment variable: ```bash sum-platform --skip-preflight backup acme SUM_SKIP_PREFLIGHT=1 sum-platform backup acme ``` ### Checks Per Command | Command | Checks | |---------|--------| | `init` | SystemConfig, DiskSpace, GiteaToken (if Gitea) | | `promote` | SiteExists, PostgresCluster, SSHConnectivity, DiskSpace, GiteaToken + GiteaRepo (if Gitea), CipherPassFile + PgBackRestStanza (if backups) | | `backup` | SiteExists, PostgresCluster, CipherPassFile, PgBackRestStanza (if backups configured) | | `restore` | SiteExists, PostgresCluster, CipherPassFile, PgBackRestStanza (if backups configured) | | `update` | SiteExists, SSHConnectivity (if `--target prod`) | | `destroy` | SiteExists (warning only) | ## Theme Management The `theme` command group provides tools for managing themes in existing projects. ### Listing Themes ```bash # List local themes (from themes/ directory or SUM_THEME_PATH) sum-platform theme list # List remote themes from sum-themes repository sum-platform theme list --remote ``` ### Checking for Updates ```bash # Compare current theme version against latest available cd /path/to/project sum-platform theme check ``` ### Viewing Available Versions ```bash # Show all available versions for a theme sum-platform theme versions theme_a ``` ### Updating Themes ```bash # Update to latest version cd /path/to/project sum-platform theme update # Update to specific version sum-platform theme update --version 1.2.0 # Allow downgrade to older version sum-platform theme update --version 1.0.0 --allow-downgrade # Force reinstall even if at target version sum-platform theme update --force ``` Theme updates are atomic with automatic rollback on failure. The lockfile at `.sum/theme.json` tracks version history. ## Site Directory Structure Each site lives at `/srv/sum/<slug>/`: ``` /srv/sum/<slug>/ ├── app/ # Django project (git checkout) ├── venv/ # Python virtualenv ├── static/ # collectstatic output ├── media/ # User uploads └── backups/ # Database backups ``` ### Ownership The CLI runs as root (`sudo sum-platform init`). After setup, `fix_site_ownership()` sets the final permissions: | Directory | Owner | Group | Why | |-----------|-------|-------|-----| | `app/` | deploy | www-data | Gunicorn reads code; deploy user runs git pulls | | `static/` | deploy | www-data | collectstatic writes here; Caddy serves directly | | `media/` | deploy | www-data | Django writes uploads here; Caddy serves directly | | `venv/` | root | root | Security — prevents application code from modifying its own runtime | | `backups/` | root | root | Security — prevents application code from accessing/tampering with backups | The `deploy` user runs gunicorn, owns application files, and is a member of `www-data`. It has no sudo access. Configured in `/etc/sum/config.yml` under `defaults.deploy_user`. Operations on site files (`app/`, `static/`, `media/`) should run as the deploy user: ```bash sudo -u deploy /srv/sum/mysite/venv/bin/python /srv/sum/mysite/app/manage.py migrate sudo -u deploy /srv/sum/mysite/venv/bin/python /srv/sum/mysite/app/manage.py collectstatic ``` Operations on `venv/` and `backups/` require root. ## Creating Sites ### With GitHub ```bash sudo sum-platform init acme --git-provider github --git-org acme-corp ``` ### With Gitea ```bash sudo sum-platform init acme --git-provider gitea --git-org clients \ --gitea-url https://gitea.agency.com ``` ### With Gitea (Custom SSH Port) ```bash sudo sum-platform init acme --git-provider gitea --git-org clients \ --gitea-url https://gitea.agency.com --gitea-ssh-port 2222 ``` ### Without Git ```bash sudo sum-platform init acme --no-git ``` ### Additional Init Options ```bash sudo sum-platform init acme --git-provider github --git-org acme-corp \ --theme theme_a \ --profile sage-stone \ --content-path /path/to/custom/content # Skip systemd service installation sudo sum-platform init acme --no-git --skip-systemd # Skip Caddy reverse proxy configuration sudo sum-platform init acme --no-git --skip-caddy # Set custom Django superuser username (default: admin) sudo sum-platform init acme --no-git --superuser myuser ``` ### Dev / Testing (Unreleased sum-core) By default, `init` installs sum-core from a pinned git tag in the boilerplate `requirements.txt`. To test with unreleased changes, use one of these overrides (listed in priority order): **`SUM_CORE_SOURCE` environment variable** — raw pip requirement line, always takes priority: ```bash SUM_CORE_SOURCE="-e /path/to/local/core" sudo sum-platform init acme --no-git SUM_CORE_SOURCE="sum-core==0.8.0" sudo sum-platform init acme --no-git ``` **`--dev` flag** — editable install from the monorepo using an absolute path (must run from within the monorepo): ```bash cd /path/to/sum-platform sudo sum-platform init acme --dev --no-git ``` **`--core-ref` flag** — install from a specific git branch or tag: ```bash sudo sum-platform init acme --core-ref develop --no-git sudo sum-platform init acme --core-ref feature/my-branch --no-git ``` `--dev` and `--core-ref` are mutually exclusive. `SUM_CORE_SOURCE` silently overrides both. ## Destroying Sites > **WARNING:** The `destroy` command is irreversible. It permanently removes a site and all its infrastructure. `destroy` reverses everything created by `init`, tearing down resources in safe order: 1. Stop and disable the systemd service 2. Remove the systemd service file and reload daemon 3. Drop the PostgreSQL cluster (and deallocate port) 4. Remove pgBackRest stanza config and backup cron 5. Remove the Caddy site config and reload 6. Remove the site directory (`/srv/sum/<slug>/`) ### Usage ```bash # Interactive — prompts you to type the site slug to confirm sudo sum-platform destroy acme # Skip confirmation prompts (for automation/scripts) sudo sum-platform destroy acme --force # Also permanently delete remote backup data sudo sum-platform destroy acme --purge-backups ``` ### Options | Option | Description | |--------|-------------| | `--force` | Skip all confirmation prompts | | `--purge-backups` | Also delete remote backup data via pgBackRest stanza-delete. **IRREVERSIBLE.** | ### Safety Features - **Slug confirmation:** You must type the exact site slug to proceed. A mismatch aborts the operation. - **Double confirmation for backups:** When `--purge-backups` is used, you must also type `DELETE BACKUPS` to confirm. - **Path safety check:** The command refuses to remove directory paths that don't contain the site slug or are too shallow (prevents accidental deletion of system directories). - **Non-fatal warnings:** If individual teardown steps fail (e.g., service already stopped), the command continues and reports warnings rather than aborting. - **Requires root:** The command escalates to root privileges automatically. ### Recommended: Back Up Before Destroying ```bash # Create a full backup first sum-platform backup acme --type=full # Then destroy (without --purge-backups to keep remote backups) sudo sum-platform destroy acme ``` ## Validating Sites The `check` command validates that a project is correctly set up and ready to run. ### Usage ```bash # Run from inside a project directory cd /srv/sum/acme/app && sum-platform check # Or specify a project name (resolves under clients/ directory) sum-platform check acme ``` ### Checks Performed | Check | What It Validates | |-------|-------------------| | Virtualenv | Virtual environment exists with required packages (django, wagtail) | | Credentials | `.env.local` exists (created during init with superuser credentials) | | Database | All migrations applied (`manage.py migrate --check`) | | Homepage | Homepage is set as the default site root page | | Theme compiled CSS | Compiled CSS exists at `theme/active/static/<slug>/css/main.css` and is >5KB | | Theme slug match | Theme slug in `.sum/theme.json` matches `theme/active/theme.json` | | Required env vars | All keys from `.env.example` are set in `.env` or environment | | sum_core import | `sum_core` package is importable in the project's Python environment | ### Example Output ``` [OK] Virtualenv: Virtualenv exists with required packages [OK] Credentials: .env.local found [OK] Database: Migrations up to date [OK] Homepage: Homepage set as site root [OK] Theme compiled CSS: /srv/sum/acme/app/theme/active/static/theme_a/css/main.css [OK] Theme slug match: theme_a [OK] Required env vars: Required env vars present [OK] sum_core import: sum_core importable ✅ All checks passed ``` Failed checks show remediation hints: ``` [FAIL] Virtualenv: Package 'django' not installed → Run 'pip install -r requirements.txt' [FAIL] Database: Pending migrations → Run 'python manage.py migrate' ``` The command exits with code `0` on success or `1` if any check fails. ## Managing Sites ```bash # Update a deployed site (pull, migrate, restart) sum-platform update acme # Update production site sum-platform update acme --target prod # Update without running migrations sum-platform update acme --skip-migrations # Backup database (PostgreSQL via pgBackRest) sum-platform backup acme --type full # Validate project setup sum-platform check acme # List available themes sum-platform theme list # List remote themes with latest versions sum-platform theme list --remote # Check for theme updates in a project cd /srv/sum/acme && sum-platform theme check # Update theme to latest version cd /srv/sum/acme && sum-platform theme update ``` ### Promoting to Production The `promote` command deploys a working staging site to a production server with a custom domain. It requires a git-enabled site (created without `--no-git`). ```bash sum-platform promote acme --domain acme-client.com ``` **Options:** | Option | Required | Description | |--------|----------|-------------| | `--domain` | Yes | Production domain for the site (e.g., `acme-client.com`) | **What promote does (11 steps):** 1. Backs up the staging database 2. Copies the backup to the production server via `scp` 3. Provisions production infrastructure (database, user, directories, `.env`) 4. Clones the site repository on production (HTTPS with token, falls back to SSH) 5. Creates a virtualenv and installs dependencies 6. Restores the database from backup 7. Syncs media files from staging via `rsync` 8. Runs Django migrations and `collectstatic` 9. Installs and enables systemd service 10. Configures Caddy reverse proxy 11. Starts the service and verifies health After completion, the command prints the production URL and reminds you to configure DNS. **Prerequisites:** - The staging site must exist and have a git repository - SSH access to the production server must be configured in `/etc/sum/config.yml` - The `production.ssh_host` setting must be set in system config ## Backup and Recovery ```bash # Create differential backup sum-platform backup acme # Create full backup sum-platform backup acme --type=full # List available backups sum-platform backup acme --list # List restore points with details sum-platform restore acme --list # Restore to specific point in time sum-platform restore acme --time "2024-01-15 14:30:00" # Restore to latest backup sum-platform restore acme --latest # Clean up pre-restore data after verifying restore sum-platform restore acme --latest --cleanup # Skip confirmation prompt (useful for automation) sum-platform restore acme --latest --confirm ``` ## Backup Monitoring The `monitor` command checks backup freshness and sends email alerts for stale backups. ```bash # Check all sites, send alerts for problems sum-platform monitor # Check without sending alerts sum-platform monitor --no-alerts # Verbose output (show all sites) sum-platform monitor -v ``` ### Cron Setup Add to crontab for hourly monitoring: ```cron 0 * * * * /usr/local/bin/sum-platform monitor ``` Alerts are sent when: - Backup is older than 48 hours - Backup status file is missing ## Configuration ### Global Config (`/etc/sum/config.yml`) Infrastructure settings only. Created via `sum-platform setup`. ```yaml agency: name: Your Agency Name staging: server: staging.example.com domain_pattern: "{slug}.staging.example.com" base_dir: /srv/sum production: server: prod.example.com ssh_host: 10.0.0.1 base_dir: /srv/sum templates: dir: /opt/your-ops/infra systemd: systemd/sum-site-gunicorn.service.template caddy: caddy/Caddyfile.template defaults: theme: theme_a seed_profile: starter deploy_user: deploy postgres_port: 5432 backups: storage_box: host: u123456.your-storagebox.de user: u123456 fingerprint: SHA256:xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx port: 23 ssh_key: /etc/sum/backup-key base_path: /backups retention: full_backups: 2 diff_backups: 7 alerts: email: alerts@agency.com ``` ### Site Config (`/srv/sum/<site>/.sum/config.yml`) Per-site configuration. Auto-created when you run `init`. **GitHub site:** ```yaml site: slug: acme theme: theme_a created: 2026-02-03T14:30:00Z git: provider: github org: acme-corp ``` **Gitea site:** ```yaml site: slug: acme theme: theme_a created: 2026-02-03T14:30:00Z git: provider: gitea org: clients url: https://gitea.agency.com ssh_port: 2222 token_env: GITEA_TOKEN ``` **No-git site:** ```yaml site: slug: acme theme: theme_a created: 2026-02-03T14:30:00Z git: null ``` ## Git Provider Setup ### GitHub Requires the GitHub CLI (`gh`) to be installed and authenticated: ```bash gh auth login ``` ### Gitea Set the API token environment variable: ```bash export GITEA_TOKEN=your-token-here ``` The CLI also detects tokens from the tea CLI config at `~/.config/tea/config.yml` as a fallback. If the token is found in the tea config, it is automatically injected into the environment for privilege escalation. **Important:** When using `--git-provider gitea`, the `init` command warns if no token is detected. The `promote` command hard-errors — Gitea access is required to clone the repository on production. Or use a custom environment variable name: ```bash sudo sum-platform init acme --git-provider gitea --git-org clients \ --gitea-url https://gitea.agency.com --gitea-token-env MY_GITEA_TOKEN ``` ## Common Workflows ### Set Up a New Client Site ```bash # 1. Create the site with git integration sudo sum-platform init acme --git-provider github --git-org acme-corp # 2. Validate the setup sum-platform check acme # 3. Visit the staging URL shown in init output ``` ### Update sum_core on a Deployed Site ```bash # 1. Create a backup before updating sum-platform backup acme --type=full # 2. Pull latest code and apply updates sum-platform update acme # 3. Verify the site is working sum-platform check acme ``` ### Recover from Backup ```bash # 1. List available restore points sum-platform restore acme --list # 2. Restore to the latest backup sum-platform restore acme --latest # 3. Verify the restored site sum-platform check acme # 4. Clean up pre-restore data once verified sum-platform restore acme --latest --cleanup ``` ### Promote a Staging Site to Production ```bash # 1. Verify the staging site is ready sum-platform check acme # 2. Create a full backup before promoting sum-platform backup acme --type=full # 3. Promote to production with a custom domain sum-platform promote acme --domain acme-client.com # 4. Configure DNS to point to the production server ``` ### Destroy a Site Safely ```bash # 1. Create a final backup (in case you need it later) sum-platform backup acme --type=full # 2. Destroy the site (keeps remote backups by default) sudo sum-platform destroy acme # 3. To also purge remote backup data (irreversible) sudo sum-platform destroy acme --purge-backups ``` ## Troubleshooting ### "Database configuration required" error The `.env` file is missing or incomplete. ```bash cp .env.example .env # Edit .env with your database credentials make db-up && make dev-reset ``` ### `check` reports "sum_core import" failure The `sum_core` package is not installed in the site's virtualenv. ```bash cd /srv/sum/acme source venv/bin/activate pip install sum-core ``` ### `destroy` warns "Could not stop service" The systemd service may already be stopped or never created (e.g., if `--skip-systemd` was used during init). This warning is non-fatal — the command continues removing other resources. ### `promote` fails with "Cannot promote without git repository" The staging site was created with `--no-git`. Promote requires a git repository to clone on the production server. Re-create the site with a git provider. ### `init` fails with permission errors The `init` command requires root. Run with `sudo`: ```bash sudo sum-platform init acme --no-git ``` ### Backup/restore fails with pgBackRest errors Ensure pgBackRest is installed and the stanza is initialized: ```bash # Check pgBackRest info for the site sudo -u postgres pgbackrest --stanza=acme info ``` If the stanza doesn't exist, it was likely not created during init. Re-running `init` is the safest fix. ## Development Install (monorepo) ```bash pip install -e ./cli ```
text/markdown
Mark Ashton
null
null
null
null
sum, cli, django, wagtail
[ "Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Programming Language :: Python :: 3.12", "Topic :: Software Development :: Build Tools", "Topic :: Software Development ...
[]
null
null
>=3.12
[]
[]
[]
[ "click>=8.1.7", "packaging>=21.0", "pyyaml>=6.0", "httpx>=0.27.0; extra == \"gitea\"" ]
[]
[]
[]
[ "Homepage, https://github.com/markashton480/sum-platform", "Repository, https://github.com/markashton480/sum-platform", "Issues, https://github.com/markashton480/sum-platform/issues", "Documentation, https://github.com/markashton480/sum-platform/tree/main/docs/dev/cli.md" ]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:21:54.828086
sum_cli-3.3.2.tar.gz
258,306
f5/b2/3b7fc8b5e8dec259a582cac69c8fad050ad095f0f63092281cf0946ae0cb/sum_cli-3.3.2.tar.gz
source
sdist
null
false
a051490070ecc0f4c671144c36d4f01d
3abecf77f18c52b6946bd53cf1c41b6f9285209db6208f21129f9c0f85ec7114
f5b23b7fc8b5e8dec259a582cac69c8fad050ad095f0f63092281cf0946ae0cb
BSD-3-Clause
[ "LICENSE" ]
248
2.4
easybib
0.4.0
Automatically fetch BibTeX entries from INSPIRE and ADS for LaTeX projects
# easybib [![Tests](https://github.com/GregoryAshton/easybib/actions/workflows/tests.yml/badge.svg)](https://github.com/GregoryAshton/easybib/actions/workflows/tests.yml) [![codecov](https://codecov.io/gh/GregoryAshton/easybib/branch/main/graph/badge.svg)](https://codecov.io/gh/GregoryAshton/easybib) Automatically fetch BibTeX entries from [INSPIRE](https://inspirehep.net/), [NASA/ADS](https://ui.adsabs.harvard.edu/), and [Semantic Scholar](https://www.semanticscholar.org/) for LaTeX projects. easybib scans your `.tex` files for citation keys, looks them up on INSPIRE, ADS, and/or Semantic Scholar, and writes a `.bib` file with the results. It handles INSPIRE texkeys (e.g. `Author:2020abc`), ADS bibcodes (e.g. `2016PhRvL.116f1102A`), and arXiv IDs (e.g. `2508.18080`). ## Installation ```bash pip install easybib ``` ## Usage ```bash easybib /path/to/latex/project easybib paper.tex ``` Pass a directory to scan all `.tex` files recursively, or a single `.tex` file. BibTeX entries are fetched and written to `references.bib`. ### Options | Flag | Description | |------|-------------| | `-o`, `--output` | Output BibTeX file (default: `references.bib`) | | `-s`, `--preferred-source` | Preferred source: `ads` (default), `inspire`, `auto`, or `semantic-scholar` | | `-a`, `--max-authors` | Truncate author lists (default: 3, use 0 for no limit) | | `-l`, `--list-keys` | List found citation keys and exit (no fetching) | | `--fresh` | Ignore existing output file and start from scratch | | `--key-type` | Enforce a single key format: `inspire`, `ads`, or `arxiv` | | `--ads-api-key` | ADS API key (overrides `ADS_API_KEY` environment variable) | | `--semantic-scholar-api-key` | Semantic Scholar API key (overrides `SEMANTIC_SCHOLAR_API_KEY` environment variable) | | `--config` | Path to config file (default: `~/.easybib.config`) | ### Examples ```bash # Scan a directory easybib ./paper --preferred-source inspire # Scan a single file easybib paper.tex # Use a custom output file easybib ./paper -o paper.bib # List citation keys without fetching easybib ./paper -l # Keep all authors easybib ./paper -a 0 ``` ### Config file You can create a config file at `~/.easybib.config` to set persistent defaults, so you don't have to pass the same flags every time: ```ini [easybib] output = references.bib max-authors = 3 preferred-source = ads ads-api-key = your-key-here semantic-scholar-api-key = your-key-here ``` All fields are optional. CLI flags override config file values, which override the built-in defaults. The `key-type` setting is also supported (see below). To use a config file at a different location: ```bash easybib ./paper --config /path/to/my.config ``` ### Source selection The `--preferred-source` flag controls where BibTeX entries are fetched from. The source determines which service provides the BibTeX data, regardless of the key format used in your `.tex` files. - **`ads`** (default) — Fetches BibTeX from ADS. If you use an INSPIRE-style key (e.g. `Author:2020abc`), easybib will cross-reference it via INSPIRE to find the corresponding ADS record, then pull the BibTeX from ADS. Falls back to INSPIRE, then Semantic Scholar, if ADS lookup fails. - **`inspire`** — Fetches BibTeX from INSPIRE. Falls back to ADS, then Semantic Scholar, if the INSPIRE lookup fails. Does not require an ADS API key unless the fallback is triggered. - **`auto`** — Chooses the source based on the key format: ADS bibcodes (e.g. `2016PhRvL.116f1102A`) are fetched from ADS, while INSPIRE-style keys are fetched from INSPIRE. Falls back to the other source, then Semantic Scholar, if the preferred one fails. - **`semantic-scholar`** — Fetches BibTeX from Semantic Scholar first, falling back to INSPIRE then ADS. Does not require an ADS API key unless the fallback is triggered. ### arXiv IDs as citation keys You can cite papers directly by their arXiv ID: ```latex \cite{2508.18080} ``` easybib fetches the BibTeX entry from your preferred source (searching by arXiv ID) and writes two entries to the `.bib` file: the full entry under its natural citation key, plus a `@misc` stub so that `\cite{2508.18080}` resolves correctly: ```bibtex @article{LIGOScientific:2025hdt, author = {Abbott, R. and others}, title = {...}, ... } @misc{2508.18080, crossref = {LIGOScientific:2025hdt} } ``` Both the new-style format (`2508.18080`) and the old-style format (`hep-ph/9905318`) are supported. ### Key type enforcement If your project uses only one type of citation key, use `--key-type` to catch accidental mixing: ```bash easybib paper.tex --key-type inspire ``` Accepted values are `inspire`, `ads`, and `arxiv`. If any key doesn't match, easybib prints the offending keys and their detected types, then exits with a non-zero status — without fetching anything: ``` Error: --key-type=inspire but 1 key(s) do not match: '2016PhRvL.116f1102A' (detected as: ads) ``` You can also set this in your config file: ```ini [easybib] key-type = inspire ``` ### Duplicate detection easybib detects when two different citation keys in your `.tex` files refer to the same paper — for example, citing both `LIGOScientific:2016aoc` and `2016PhRvL.116f1102A`. Detection is based on: - The citation key returned by the API (before any key replacement) - The arXiv eprint ID in the BibTeX entry - The DOI in the BibTeX entry When a duplicate is found, the second entry is skipped and a warning is printed at the end of the run: ``` Warning: 1 key(s) skipped — they refer to the same paper as an earlier key. Please use a single key per paper in your .tex files: '2016PhRvL.116f1102A' duplicates 'LIGOScientific:2016aoc' (source key 'LIGOScientific:2016aoc') ``` ### ADS API key When using ADS as the source (the default), provide your API key either via the command line: ```bash easybib ./paper --ads-api-key your-key-here ``` Or as an environment variable: ```bash export ADS_API_KEY="your-key-here" ``` Get a key from https://ui.adsabs.harvard.edu/user/settings/token. ### Semantic Scholar API key Semantic Scholar's API works without a key but is rate-limited. For heavier use, provide an API key either via the command line: ```bash easybib ./paper --semantic-scholar-api-key your-key-here ``` Or as an environment variable: ```bash export SEMANTIC_SCHOLAR_API_KEY="your-key-here" ``` Get a key from https://www.semanticscholar.org/product/api. ## How it works 1. Scans `.tex` files for `\cite{...}`, `\citep{...}`, `\citet{...}`, and related commands 2. Accepts INSPIRE texkeys (`Author:2020abc`), ADS bibcodes (`2016PhRvL.116f1102A`), and arXiv IDs (`2508.18080` or `hep-ph/9905318`); warns and skips anything else 3. Optionally enforces that all keys are of a single type (`--key-type`) 4. Fetches BibTeX from the preferred source, with automatic fallback 5. For INSPIRE/ADS keys: replaces the citation key to match what is in your `.tex` file 6. For arXiv IDs: keeps the entry's natural key and appends a `@misc` crossref stub so `\cite{arxiv_id}` resolves correctly 7. Detects duplicate entries (same paper cited under different keys) and skips them with a warning 8. Truncates long author lists 9. Skips keys already present in the output file (use `--fresh` to override)
text/markdown
null
Gregory Ashton <gregory.ashton@ligo.org>
null
null
null
null
[]
[]
null
null
>=3.9
[]
[]
[]
[ "requests", "pytest; extra == \"test\"", "pytest-cov; extra == \"test\"" ]
[]
[]
[]
[ "Homepage, https://github.com/GregoryAshton/easybib", "Repository, https://github.com/GregoryAshton/easybib" ]
twine/6.2.0 CPython/3.12.2
2026-02-19T12:20:35.352854
easybib-0.4.0.tar.gz
21,814
10/d6/e5ac0b73728e00f65c65260c2808a82b717828045b70fb71c80d24b8a8d0/easybib-0.4.0.tar.gz
source
sdist
null
false
04d396e6949589de94e7a1c1d4ffeeb9
af1c0b90f80b586cb60a4b9c69675d0993ee20c8371c0d33a17cb97a59591a07
10d6e5ac0b73728e00f65c65260c2808a82b717828045b70fb71c80d24b8a8d0
MIT
[]
249
2.4
uipath-core
0.5.1
UiPath Core abstractions
# UiPath Core [![PyPI - Version](https://img.shields.io/pypi/v/uipath-core)](https://pypi.org/project/uipath-core/) [![PyPI downloads](https://img.shields.io/pypi/dm/uipath-core.svg)](https://pypi.org/project/uipath-core/) [![Python versions](https://img.shields.io/pypi/pyversions/uipath-core.svg)](https://pypi.org/project/uipath-core/) Core abstractions and contracts for the UiPath Python SDK. ## Installation ```bash pip install uipath-core ``` ## Modules ### Errors Exception hierarchy for UiPath trigger errors with category-based classification. - **`ErrorCategory`**: Enum: `DEPLOYMENT`, `SYSTEM`, `UNKNOWN`, `USER` - **`UiPathFaultedTriggerError`**: Base trigger error with category and detail - **`UiPathPendingTriggerError`**: Pending trigger variant ```python from uipath.core.errors import ErrorCategory, UiPathFaultedTriggerError ``` ### Serialization JSON serialization utilities for complex Python types. Handles Pydantic models (v1 & v2), dataclasses, enums, datetime/timezone objects, sets, tuples, and named tuples. - **`serialize_json(obj)`**: Serialize any object to a JSON string - **`serialize_defaults(obj)`**: Custom `default` handler for `json.dumps()` ```python from uipath.core.serialization import serialize_json ``` ### Tracing OpenTelemetry integration with UiPath execution tracking. Provides function instrumentation, span lifecycle management, custom exporters, and batch/simple span processors with automatic `execution.id` propagation. - **`@traced`**: Decorator for sync/async function instrumentation. Supports custom span names, run types, input/output processors, and non-recording spans - **`UiPathTraceManager`**: Manages `TracerProvider`, span exporters, and processors. Provides `start_execution_span()` context manager and span retrieval by execution ID - **`UiPathSpanUtils`**: Span registry and parent context management - **`UiPathTraceSettings`**: Configuration model with optional span filtering ```python from uipath.core.tracing import traced, UiPathTraceManager @traced(name="my_operation", run_type="tool") def do_work(input: str) -> str: return process(input) ``` ### Guardrails Deterministic rule-based validation for inputs and outputs. Rules are evaluated pre-execution (input-only) and post-execution (all rules), with flexible field selection using dot-notation paths and array access (`[*]`). **Rule types:** - **`WordRule`**: String pattern matching - **`NumberRule`**: Numeric constraint validation - **`BooleanRule`**: Boolean assertions - **`UniversalRule`**: Always-apply constraints **Field selection:** - **`AllFieldsSelector`**: Apply to all fields of a given source (input/output) - **`SpecificFieldsSelector`**: Target specific fields by path **Service:** - **`DeterministicGuardrailsService`**: Evaluates guardrail rules against inputs/outputs, returning `GuardrailValidationResult` with pass/fail status and reason ```python from uipath.core.guardrails import DeterministicGuardrailsService, GuardrailValidationResultType ``` ### Chat Pydantic models for the UiPath conversation event protocol. Defines the streaming event schema between clients and LLM/agent backends. **Hierarchy:** ``` Conversation → Exchange → Message → Content Parts (with Citations) → Tool Calls (with Results) → Interrupts (human-in-the-loop) ``` Supports session capabilities negotiation, async input streams (audio/video), tool call confirmation interrupts, URL and media citations, and inline/external value references. ```python from uipath.core.chat import UiPathConversationEvent, UiPathSessionStartEvent ``` ## Dependencies | Package | Version | |---|---| | `pydantic` | `>=2.12.5, <3` | | `opentelemetry-sdk` | `>=1.39.0, <2` | | `opentelemetry-instrumentation` | `>=0.60b0, <1` |
text/markdown
null
null
null
Marius Cosareanu <marius.cosareanu@uipath.com>, Cristian Pufu <cristian.pufu@uipath.com>
null
null
[ "Intended Audience :: Developers", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Software Development :: Build Tools" ]
[]
null
null
>=3.11
[]
[]
[]
[ "opentelemetry-instrumentation<1.0.0,>=0.60b0", "opentelemetry-sdk<2.0.0,>=1.39.0", "pydantic<3.0.0,>=2.12.5" ]
[]
[]
[]
[ "Homepage, https://uipath.com", "Repository, https://github.com/UiPath/uipath-core-python", "Documentation, https://uipath.github.io/uipath-python/" ]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:20:14.295586
uipath_core-0.5.1.tar.gz
116,874
71/13/7c3a3982b714dcb6edf7b563fe2a01eb26903db2e7a232e71a8d074cf39a/uipath_core-0.5.1.tar.gz
source
sdist
null
false
f1cea23189408ac53b13e2f0633c9848
63afb06ce2e6e48cb27469576fc1c9606e544baacb61f92b8352473031d2f189
71137c3a3982b714dcb6edf7b563fe2a01eb26903db2e7a232e71a8d074cf39a
null
[ "LICENSE" ]
32,599
2.4
e-invoice-mcp
0.1.21
MCP Server for Fintom8 E-Invoicing services (Conversion, Validation, Correction)
# Fintom8 E-Invoicing Agent (MCP) [![PyPI version](https://img.shields.io/pypi/v/e-invoice-mcp.svg)](https://pypi.org/project/e-invoice-mcp/) [![MCP Certified](https://img.shields.io/badge/MCP-Certified-blue)](https://mcpmarket.com) [![Peppol Ready](https://img.shields.io/badge/Peppol-3.0.20-green)](https://fintom8.com) **The official Model Context Protocol (MCP) server for Fintom8.** This server acts as an intelligent bridge to the AI-driven Fintom8 E-Invoice Platform. It enables autonomous agents to validate, audit, and correct e-invoices against the latest European standards (EN16931) and Peppol regulations. --- ## 🚀 Features - **PDF to UBL Conversion:** Generate compliant e-invoices from any format, including PDF, XML, JSON and CSV. - **Automated Validation:** Validate your ZUGFeRD and UBL e-invoices against 300+ EN16931 rules. AI-driven PDF–XML comparison ensures your invoices meet compliance standards. - **Correction:** Automatically correct errors in XML files to ensure seamless integration with your system. --- ## 🛠️ Installation The easiest way to use the server is to install it via pip: ```bash pip install e-invoice-mcp ``` ### Run the Server Once installed, you can start the server with: ```bash e-invoice-mcp ``` --- ## 🔑 AI Client Configuration ### Claude Desktop (Action required) To use these tools in Claude, add the following to your configuration file: **File:** `~/Library/Application Support/Claude/claude_desktop_config.json` ```json { "mcpServers": { "fintom8": { "command": "e-invoice-mcp" } } } ``` --- ## 📦 Included Tools ### 1. `convert_pdf_to_invoice` Converts PDF invoices to structured UBL format. - **Args**: `pdf_path` (path). - **Output**: UBL XML. ### 2. `validate_invoice` (Basic Validation) Validates UBL/Peppol XML invoices against compliance rules. - **Args**: `xml_content` (string) or `xml_path` (path). - **Output**: Simple JSON report (is_valid, errors). ### 3. `validate_invoice_v2` (Advanced Validation) Deep validation with optional AI explanations. - **Args**: `xml_content` (string) or `xml_path` (path). - **Output**: Detailed compliance report. ### 4. `correct_invoice_xml` AI-powered correction of invalid XML invoices. - **Args**: `xml_content` (string) or `xml_path` (path). - **Output**: Fixed XML content. --- ## � Privacy & Security This server acts as a thin client proxy. Data is processed on secure Fintom8 production servers and is not used for AI model training. **License:** MIT **Website:** [fintom8.com](https://fintom8.com)
text/markdown
null
Igor Nikolaienko <igor@fintom8.com>
null
null
null
null
[]
[]
null
null
>=3.9
[]
[]
[]
[ "fastmcp", "httpx" ]
[]
[]
[]
[ "Homepage, https://github.com/NikolaienkoIgor/Fintom8-E-invoice-MCP-server", "Issues, https://github.com/NikolaienkoIgor/Fintom8-E-invoice-MCP-server/issues" ]
twine/6.2.0 CPython/3.10.10
2026-02-19T12:20:12.421442
e_invoice_mcp-0.1.21.tar.gz
5,318
e4/a0/07b35a21192477ae12ed8fcb2af4a4faff3385a1e5fe8eeb923119cc8663/e_invoice_mcp-0.1.21.tar.gz
source
sdist
null
false
dcad9ce8f3fc404e4acc2097e37c29c2
43ce6f9b3e7d02710f8ca9e5e1eda0bd5fcd71e15d33fef0f03ae476c23be0ce
e4a007b35a21192477ae12ed8fcb2af4a4faff3385a1e5fe8eeb923119cc8663
null
[ "LICENSE" ]
253
2.4
aiohomematic-config
2026.2.7
Presentation-layer library for Homematic device configuration UI.
# aiohomematic-config [![CI](https://github.com/sukramj/aiohomematic-config/actions/workflows/test-run.yaml/badge.svg)](https://github.com/sukramj/aiohomematic-config/actions/workflows/test-run.yaml) [![codecov](https://codecov.io/gh/sukramj/aiohomematic-config/branch/devel/graph/badge.svg)](https://codecov.io/gh/sukramj/aiohomematic-config) [![PyPI](https://img.shields.io/pypi/v/aiohomematic-config.svg)](https://pypi.org/project/aiohomematic-config/) Presentation-layer library for Homematic device configuration UI. Transforms Homematic device paramset descriptions into UI-optimized structures. No RPC knowledge, no CCU access -- operates purely on data structures from [aiohomematic](https://github.com/sukramj/aiohomematic). ## Installation ```bash pip install aiohomematic-config ``` ## Quick Start ```python from aiohomematic_config import FormSchemaGenerator generator = FormSchemaGenerator(locale="en") schema = generator.generate( descriptions=descriptions, current_values=current_values, channel_type="HEATING_CLIMATECONTROL_TRANSCEIVER", ) # schema is a Pydantic model, JSON-serializable print(schema.model_dump_json(indent=2)) ``` ## Key Components | Component | Purpose | | --------------------- | ------------------------------------------------ | | `FormSchemaGenerator` | ParameterData + values -> JSON form schemas | | `ParameterGrouper` | Flat parameter list -> grouped sections | | `LabelResolver` | Technical parameter IDs -> human-readable labels | | `ConfigSession` | Change tracking, undo/redo, dirty state | | `ConfigExporter` | Serialize/deserialize device configurations | | `WidgetType` mapping | ParameterType -> appropriate UI widget | ## License MIT License - see [LICENSE](LICENSE) for details.
text/markdown
null
SukramJ <sukramj@icloud.com>
null
null
MIT License
home, automation, homematic, configuration, paramset
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Natural Language :: German", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Lan...
[]
null
null
>=3.13
[]
[]
[]
[ "aiohomematic>=2026.2.20", "pydantic>=2.10.0" ]
[]
[]
[]
[ "Homepage, https://github.com/sukramj/aiohomematic-config", "Source Code, https://github.com/sukramj/aiohomematic-config", "Bug Reports, https://github.com/sukramj/aiohomematic-config/issues", "Changelog, https://github.com/sukramj/aiohomematic-config/blob/devel/changelog.md" ]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:19:58.374689
aiohomematic_config-2026.2.7.tar.gz
275,176
cc/78/a656635afb186d5a756ef04bfc8d9053d3f84142b0d49a4db1e854c1123a/aiohomematic_config-2026.2.7.tar.gz
source
sdist
null
false
02a071414d0702e197827de71c5bd847
2e13c06bc41e591d7ff83452863e05cd2cb37bb92919aa363c1c25ab15a21fb8
cc78a656635afb186d5a756ef04bfc8d9053d3f84142b0d49a4db1e854c1123a
null
[ "LICENSE" ]
695
2.4
django-generic-api-permissions
3.0.1
Generic API permissions and visibilities for Django
# Django Generic API Permissions Django Generic API Permissions (DGAP) is a framework to make your Django Rest Framework API user-extensible for common use cases. Specifically, it provides a simple API for your users to define specific visibilities, permissions, and validations. Assume you have an API service that implements a blogging backend. You have a `Post`, and a `Comment` model. When deployed as a public blog, you want admins users to be able to post, and authenticated users be able to comment. Everybody is allowed to read posts and comments. But the same software should also be used in a company-internal deployment where the rules are different: Anonymous users should not see anything, and you have specific rules as to who can write posts. DGAP makes it easy to implement the blog once, and make the permissions, visibilities, and validations custom to each deployment. ## Concepts DGAP provides you with three configuration settings: Visibilities, Permissions, and Validations. - The **visibilities** are run when getting data from the API. They define, on a per-user base, who can see which data. - The **validations** are run on create and update operations, so data can be checked and modified before the update takes place. - The **permissions** then define what a user can do with a given (visible) piece of data. ## Installation for app developers If you want to integrate DGAP into your app, these are the steps you need. Install DGAP (and add to your requirements files etc) first. ```bash pip install django-generic-api-permissions ``` Then, add `generic_permissions.apps.GenericPermissionsConfig` to your `INSTALLED_APPS`: ```python INSTALLED_APPS = ( ... "generic_permissions.apps.GenericPermissionsConfig", ... ) ``` ### Visibility subsystem The visibility part defines what you can see at all. Anything you cannot see, you're implicitly also not allowed to modify. The visibility classes define what you see depending on your roles, permissions, etc. Building on top of this follow the permission classes (see below) that define what you can do with the data you see. For the visibilities, extend your DRF `ViewSet` classes with the `VisibilityViewMixin`: ```python # views.py from rest_framework.viewsets import ModelViewSet from generic_permissions.visibilities import VisibilityViewMixin class MyModelViewset(VisibilityViewMixin, ModelViewSet): serializer_class = MyModelSerializers queryset = ... ``` Data leaks in REST Framework may happen if you use includes (or even only references) from a model the user may see to something that should be hidden. To avoid such leaks, make sure to use a subclassed related field (either by creating your own using the provided `VisibilityRelatedFieldMixin`, or by using one of the provided types. See example below). To set it for every relation in the serializer use DRFs `serializer_related_field` attribute in the serializer. It's important to be aware of potential issues when updating (PATCH) existing relationships, especially when some relationships are hidden due to visibility settings. If not handled correctly, the hidden relationships may be unintentionally removed during an update, resulting in only the new relationships being set. To avoid this, you must incorporate the `VisibilitySerializerMixin` into your serializer where you're using the `VisibilityRelatedFieldMixin` for the relationship field. This ensures that hidden relationships are properly accounted for during updates. Remember to define the `VisibilitySerializerMixin` after the `ValidatorMixin`. This order is crucial because it ensures that validations are performed first, and only then are the relationships updated. This step is vital to maintain the integrity of your data and prevent accidental loss of hidden relationships. ```python # serializers.py from rest_framework.viewsets import ModelSerializer from generic_permissions.visibilities import VisibilityPrimaryKeyRelatedField, VisibilitySerializerMixin class MyModelSerializers(VisibilitySerializerMixin, ModelSerializer): serializer_related_field = VisibilityPrimaryKeyRelatedField ``` A few subclassed fields are provided for different types of `RelatedField`: - `VisibilityPrimaryKeyRelatedField` - `VisibilityResourceRelatedField` - `VisibilitySerializerMethodResourceRelatedField` If a different relation field variation is needed extend it with `VisibilityRelatedFieldMixin`: ```python from generic_permissions.visibilities import VisibilityRelatedFieldMixin class CustomRelationField(VisibilityRelatedFieldMixin): pass ``` #### Bypassing visibilities for foreign keys for 1:n and n:m DGAP allows you to enforce visibility checks on foreign keys as well. Sometimes, you might want to bypass this, as it's not always necessary. For example, if you have a "document" with multiple "versions" as a 1:n relationship, you don't want to filter the versions queryset again, as it conforms to the same rules as the documents (that you 've already filtered) To configure this, add the following key to your `settings.py`: ```python GENERIC_PERMISSIONS_BYPASS_VISIBILITIES = { "my_app.Document": ["versions"], # attribute name as defined on the model "my_app.File": "__all__" # for all fields } ``` Only use `__all__` for testing or if you're sure that you are not accidentally exposing any related models. ### Permission subsystem Similarly, for the permissions system, add the `PermissionViewMixin` to your views: ```python # views.py from rest_framework.viewsets import ModelViewSet from generic_permissions.permissions import PermissionViewMixin class MyModelViewset(PermissionViewMixin, VisibilityViewMixin, ModelViewSet): serializer_class = ... queryset = ... ``` You may use only one of the two mixins, or both, depending on your needs. ### Validation subsystem Last, for the validation system, you extend your **serializer** with a mixin: ```python # serializers.py from rest_framework.serializers import ModelSerializer from generic_permissions.serializers import PermissionSerializerMixin from generic_permissions.validation import ValidatorMixin from myapp import models class MyModelSerializer(ValidatorMixin, ModelSerializer): # my field definitions... class Meta: model = models.MyModel fields = "__all__" ``` ## Usage - for people deploying a DGAP-equipped app Say you have an blog you want to deploy that uses DGAP. You want public blog posts, but the comment section should only be visible for authenticated users. For this, you would define a custom visibility class that limits access accordingly. ### Visibilities First, let's define the visibility class: ```python # my_custom_visibilities.py from generic_permissions.visibilities import filter_queryset_for from my_app.models import Post, Comment class CustomVisibility: @filter_queryset_for(Post) def filter_posts(self, queryset, request): # no filtering on blog posts return queryset @filter_queryset_for(Comment) def filter_comments(self, queryset, request): # Only authenticated users shall see comments if request.user.is_authenticated: return queryset else: return queryset.none() ``` Once done, open `settings.py` and point the `GENERIC_PERMISSIONS_VISIBILITY_CLASSES` setting to the class you just defined. It is a list of strings that name the visibility classes. ```python GENERIC_PERMISSIONS_VISIBILITY_CLASSES = ['my_custom_visibilities.CustomVisibility'] ``` Note: The setting may be defined using env variables depending on the project. In that case, set the value via that way instead. Some times, you have visibilities that you want to combine: Say one visibility class provides read access for user group A, another class provides access for user group B. You want to combine those in a simple way. For this, we have provided you the `Union` visibility: ```python from generic_permissions.visibilities import Union class MyFirstVisibility: # ... class MySecondVisibility: # ... class ResultingVisibility(Union): # Define a property `visibility_classes`. Those # will then be checked both, and if either one allows # an object to be seen, it will be visible to the user. visibility_classes = [MyFirstVisibility, MySecondVisibility] ``` ### Permissions Permission classes define who may perform which data mutation. They can be configured via `GENERIC_PERMISSIONS_PERMISSION_CLASSES`. To write custom permission classes, you create a simple class, and decorate the methods that define the permissions accordingly. There are two types of methods in the permissions system: - `permission_for`: Marks methods that define generic access permissions for a given model. They are always checked first. Those methods will receive one positional argument, namely the `request` object and one named argument `action` which contains the current [DRF view action][drf-actions] as string. - `object_permission_for`: Define whether access to a specific object shall be granted. This called for all other operations **except** creation. These methods will receive two positional arguments: First, the `request` object, and second, the model instance that is being accessed in the request. In addition to that, it will receive one named argument `action` which contains the current [DRF view action][drf-actions] as string. [drf-actions]: https://www.django-rest-framework.org/api-guide/viewsets/#viewset-actions The following example carries on the Blog concept from above. We want only admins to edit/update blog posts, and authenticated users to comment. Nobody should be able to edit their comments. We also show the concept of combining two permission classes here. DGAP looks at the whole inheritance tree to figure out the permissions, so you can leverage that to avoid code duplication. You can find more information about the `request` object in the [Django documentation](https://docs.djangoproject.com/en/3.1/ref/request-response/#httprequest-objects) ```python from generic_permissions.permissions import permission_for, object_permission_for from my_app.models import Post, Comment class OnlyAuthenticated: @permission_for(object) def has_permission_default(self, request, *args, **kwargs): # No permission is granted for any non-authenticated users return request.user.is_authenticated class BlogPermissions: @permission_for(Comment) def has_permission_for_comment(self, request, *args, **kwargs): # comments can be added, but not updated return request.method == "POST" @permission_for(Post) def has_permission_for_post(self, request, instance, *args, **kwargs): # Only admins can work on Posts return "admin" in request.user.groups @object_permission_for(Post) def has_object_permission_for_post(self, request, instance, *args, **kwargs): # Of the admins, changing a Post is only allowed to the author. return instance.author == request.user ``` The following pre-defined classes are available: - `generic_permissions.permissions.AllowAny`: allow any users to perform any mutation (default) - `generic_permissions.permissions.DenyAll`: deny all operations to any object. You can use this as a base class for your permissions - as long as you don't allow something, it will be denied. ### Data validation Once the permission to access or modify an object is granted, you may want to apply some custom validation as well. In the example we're using here, we assume some user registration form. We want to ensure that the username contains only lowercase letters. For this, you can use the `GENERIC_PERMISSIONS_VALIDATION_CLASSES` setting. The settings is a list of strings, representing a list of class names). Here's an example validator class that ensures the username is lower case. ```python from generic_permissions.validation import validator_for from my_app.models import User class LowercaseUsername: @validator_for(User) def lowercase_username(self, data, context): data["username"] = data["username"].lower() return data ``` The `@validator_for` decorator tells DGAP that the method shall be called when a `User` is modified. The data passed in is already parsed and validated by the REST framework, and it is expected that the method returns a `dict` with a compatible structure. You may also `raise ValidationError("some message")` if you don't want the validation to succeed. The second parameter, `context`, is a `dict` containing the DRF context: Access `context['request']` to get the request (if validation depends on the user, for example).
text/markdown
Adfinis
info@adfinis.com
null
null
null
null
[ "Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: Django", "Framework :: Django :: 3.2", "Framework :: Django :: 4.2", "Intended Audience :: Developers", "License :: OSI Approved :: GNU Lesser General Public License v3 or later (LGPLv3+)", "Operating System :: OS Indepe...
[]
null
null
<4,>=3.12
[]
[]
[]
[ "django>=3.2", "djangorestframework<4.0,>=3.14" ]
[]
[]
[]
[ "Repository, https://github.com/adfinis/django-generic-api-permissions" ]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:18:48.813432
django_generic_api_permissions-3.0.1.tar.gz
22,472
f7/75/4fadafd40ce5b43cb676d1920454cef70ef62537f21f36a7dc88d0e44c87/django_generic_api_permissions-3.0.1.tar.gz
source
sdist
null
false
d6a5207ed399cd1717cdb0025f97a881
9cfea423f023343b086e9f60750baf6a116d09e5ff5c6c7d8974f78368590240
f7754fadafd40ce5b43cb676d1920454cef70ef62537f21f36a7dc88d0e44c87
null
[ "LICENSE" ]
337
2.4
mcp-email-server
0.6.1
IMAP and SMTP via MCP Server
# mcp-email-server [![Release](https://img.shields.io/github/v/release/ai-zerolab/mcp-email-server)](https://img.shields.io/github/v/release/ai-zerolab/mcp-email-server) [![Build status](https://img.shields.io/github/actions/workflow/status/ai-zerolab/mcp-email-server/main.yml?branch=main)](https://github.com/ai-zerolab/mcp-email-server/actions/workflows/main.yml?query=branch%3Amain) [![codecov](https://codecov.io/gh/ai-zerolab/mcp-email-server/branch/main/graph/badge.svg)](https://codecov.io/gh/ai-zerolab/mcp-email-server) [![Commit activity](https://img.shields.io/github/commit-activity/m/ai-zerolab/mcp-email-server)](https://img.shields.io/github/commit-activity/m/ai-zerolab/mcp-email-server) [![License](https://img.shields.io/github/license/ai-zerolab/mcp-email-server)](https://img.shields.io/github/license/ai-zerolab/mcp-email-server) [![smithery badge](https://smithery.ai/badge/@ai-zerolab/mcp-email-server)](https://smithery.ai/server/@ai-zerolab/mcp-email-server) IMAP and SMTP via MCP Server - **Github repository**: <https://github.com/ai-zerolab/mcp-email-server/> - **Documentation** <https://ai-zerolab.github.io/mcp-email-server/> ## Installation ### Manual Installation We recommend using [uv](https://github.com/astral-sh/uv) to manage your environment. Try `uvx mcp-email-server@latest ui` to config, and use following configuration for mcp client: ```json { "mcpServers": { "zerolib-email": { "command": "uvx", "args": ["mcp-email-server@latest", "stdio"] } } } ``` This package is available on PyPI, so you can install it using `pip install mcp-email-server` After that, configure your email server using the ui: `mcp-email-server ui` ### Environment Variable Configuration You can also configure the email server using environment variables, which is particularly useful for CI/CD environments like Jenkins. zerolib-email supports both UI configuration (via TOML file) and environment variables, with environment variables taking precedence. ```json { "mcpServers": { "zerolib-email": { "command": "uvx", "args": ["mcp-email-server@latest", "stdio"], "env": { "MCP_EMAIL_SERVER_ACCOUNT_NAME": "work", "MCP_EMAIL_SERVER_FULL_NAME": "John Doe", "MCP_EMAIL_SERVER_EMAIL_ADDRESS": "john@example.com", "MCP_EMAIL_SERVER_USER_NAME": "john@example.com", "MCP_EMAIL_SERVER_PASSWORD": "your_password", "MCP_EMAIL_SERVER_IMAP_HOST": "imap.gmail.com", "MCP_EMAIL_SERVER_IMAP_PORT": "993", "MCP_EMAIL_SERVER_SMTP_HOST": "smtp.gmail.com", "MCP_EMAIL_SERVER_SMTP_PORT": "465" } } } } ``` #### Available Environment Variables | Variable | Description | Default | Required | | --------------------------------------------- | ------------------------------------------------------ | ------------- | -------- | | `MCP_EMAIL_SERVER_ACCOUNT_NAME` | Account identifier | `"default"` | No | | `MCP_EMAIL_SERVER_FULL_NAME` | Display name | Email prefix | No | | `MCP_EMAIL_SERVER_EMAIL_ADDRESS` | Email address | - | Yes | | `MCP_EMAIL_SERVER_USER_NAME` | Login username | Same as email | No | | `MCP_EMAIL_SERVER_PASSWORD` | Email password | - | Yes | | `MCP_EMAIL_SERVER_IMAP_HOST` | IMAP server host | - | Yes | | `MCP_EMAIL_SERVER_IMAP_PORT` | IMAP server port | `993` | No | | `MCP_EMAIL_SERVER_IMAP_SSL` | Enable IMAP SSL | `true` | No | | `MCP_EMAIL_SERVER_IMAP_VERIFY_SSL` | Verify IMAP SSL certificates (disable for self-signed) | `true` | No | | `MCP_EMAIL_SERVER_SMTP_HOST` | SMTP server host | - | Yes | | `MCP_EMAIL_SERVER_SMTP_PORT` | SMTP server port | `465` | No | | `MCP_EMAIL_SERVER_SMTP_SSL` | Enable SMTP SSL | `true` | No | | `MCP_EMAIL_SERVER_SMTP_START_SSL` | Enable STARTTLS | `false` | No | | `MCP_EMAIL_SERVER_SMTP_VERIFY_SSL` | Verify SSL certificates (disable for self-signed) | `true` | No | | `MCP_EMAIL_SERVER_ENABLE_ATTACHMENT_DOWNLOAD` | Enable attachment download | `false` | No | | `MCP_EMAIL_SERVER_SAVE_TO_SENT` | Save sent emails to IMAP Sent folder | `true` | No | | `MCP_EMAIL_SERVER_SENT_FOLDER_NAME` | Custom Sent folder name (auto-detect if not set) | - | No | ### Enabling Attachment Downloads By default, downloading email attachments is disabled for security reasons. To enable this feature, you can either: **Option 1: Environment Variable** ```json { "mcpServers": { "zerolib-email": { "command": "uvx", "args": ["mcp-email-server@latest", "stdio"], "env": { "MCP_EMAIL_SERVER_ENABLE_ATTACHMENT_DOWNLOAD": "true" } } } } ``` **Option 2: TOML Configuration** Add `enable_attachment_download = true` to your TOML configuration file (`~/.config/zerolib/mcp_email_server/config.toml`): ```toml enable_attachment_download = true [[emails]] # ... your email configuration ``` Once enabled, you can use the `download_attachment` tool to save email attachments to a specified path. ### Saving Sent Emails to IMAP Sent Folder By default, sent emails are automatically saved to your IMAP Sent folder. This ensures that emails sent via the MCP server appear in your email client (Thunderbird, webmail, etc.). The server auto-detects common Sent folder names: `Sent`, `INBOX.Sent`, `Sent Items`, `Sent Mail`, `[Gmail]/Sent Mail`. **To specify a custom Sent folder name** (useful for providers with non-standard folder names): **Option 1: Environment Variable** ```json { "mcpServers": { "zerolib-email": { "command": "uvx", "args": ["mcp-email-server@latest", "stdio"], "env": { "MCP_EMAIL_SERVER_SENT_FOLDER_NAME": "INBOX.Sent" } } } } ``` **Option 2: TOML Configuration** ```toml [[emails]] account_name = "work" save_to_sent = true sent_folder_name = "INBOX.Sent" # ... rest of your email configuration ``` **To disable saving to Sent folder**, set `MCP_EMAIL_SERVER_SAVE_TO_SENT=false` or `save_to_sent = false` in your TOML config. ### Self-Signed Certificates (e.g., ProtonMail Bridge) If you're using a local mail server with self-signed certificates (like ProtonMail Bridge), you'll need to disable SSL certificate verification: ```json { "mcpServers": { "zerolib-email": { "command": "uvx", "args": ["mcp-email-server@latest", "stdio"], "env": { "MCP_EMAIL_SERVER_IMAP_VERIFY_SSL": "false", "MCP_EMAIL_SERVER_SMTP_VERIFY_SSL": "false" } } } } ``` Or in TOML configuration: ```toml [[emails]] account_name = "protonmail" # ... other settings ... [emails.incoming] verify_ssl = false [emails.outgoing] verify_ssl = false ``` For separate IMAP/SMTP credentials, you can also use: - `MCP_EMAIL_SERVER_IMAP_USER_NAME` / `MCP_EMAIL_SERVER_IMAP_PASSWORD` - `MCP_EMAIL_SERVER_SMTP_USER_NAME` / `MCP_EMAIL_SERVER_SMTP_PASSWORD` Then you can try it in [Claude Desktop](https://claude.ai/download). If you want to intergrate it with other mcp client, run `$which mcp-email-server` for the path and configure it in your client like: ```json { "mcpServers": { "zerolib-email": { "command": "{{ ENTRYPOINT }}", "args": ["stdio"] } } } ``` If `docker` is avaliable, you can try use docker image, but you may need to config it in your client using `tools` via `MCP`. The default config path is `~/.config/zerolib/mcp_email_server/config.toml` ```json { "mcpServers": { "zerolib-email": { "command": "docker", "args": ["run", "-it", "ghcr.io/ai-zerolab/mcp-email-server:latest"] } } } ``` ### Installing via Smithery To install Email Server for Claude Desktop automatically via [Smithery](https://smithery.ai/server/@ai-zerolab/mcp-email-server): ```bash npx -y @smithery/cli install @ai-zerolab/mcp-email-server --client claude ``` ## Usage ### Replying to Emails To reply to an email with proper threading (so it appears in the same conversation in email clients): 1. First, fetch the original email to get its `message_id`: ```python emails = await get_emails_content(account_name="work", email_ids=["123"]) original = emails.emails[0] ``` 2. Send your reply using `in_reply_to` and `references`: ```python await send_email( account_name="work", recipients=[original.sender], subject=f"Re: {original.subject}", body="Thank you for your email...", in_reply_to=original.message_id, references=original.message_id, ) ``` The `in_reply_to` parameter sets the `In-Reply-To` header, and `references` sets the `References` header. Both are used by email clients to thread conversations properly. ## Development This project is managed using [uv](https://github.com/ai-zerolab/uv). Try `make install` to install the virtual environment and install the pre-commit hooks. Use `uv run mcp-email-server` for local development. ## Releasing a new version - Create an API Token on [PyPI](https://pypi.org/). - Add the API Token to your projects secrets with the name `PYPI_TOKEN` by visiting [this page](https://github.com/ai-zerolab/mcp-email-server/settings/secrets/actions/new). - Create a [new release](https://github.com/ai-zerolab/mcp-email-server/releases/new) on Github. - Create a new tag in the form `*.*.*`. For more details, see [here](https://fpgmaas.github.io/cookiecutter-uv/features/cicd/#how-to-trigger-a-release).
text/markdown
null
ai-zerolab <jizhongsheng957@gmail.com>
null
null
null
IMAP, MCP, SMTP, email
[ "Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: P...
[]
null
null
<4.0,>=3.10
[]
[]
[]
[ "aioimaplib>=2.0.1", "aiosmtplib>=4.0.0", "gradio>=6.0.1", "jinja2>=3.1.5", "loguru>=0.7.3", "mcp[cli]<2,>=1.23.0", "pydantic-settings[toml]>=2.11.0", "pydantic>=2.11.0", "tomli-w>=1.2.0", "typer>=0.15.1" ]
[]
[]
[]
[ "Homepage, https://ai-zerolab.github.io/mcp-email-server/", "Repository, https://github.com/ai-zerolab/mcp-email-server", "Documentation, https://ai-zerolab.github.io/mcp-email-server/" ]
uv/0.6.2
2026-02-19T12:16:54.965594
mcp_email_server-0.6.1.tar.gz
199,227
dc/6f/a4380e9fbfffb2b84a29a321c178f6ec8c48c092e96f5a4fab705aa84b07/mcp_email_server-0.6.1.tar.gz
source
sdist
null
false
ac8b036f6667c7a262915fb111399881
6fe6cdf6a45b975593ee7111e1ce6889e0ed56711a4221090ed1a66e50ffaa70
dc6fa4380e9fbfffb2b84a29a321c178f6ec8c48c092e96f5a4fab705aa84b07
null
[ "LICENSE" ]
700
2.4
venturalitica
0.5.0
Frictionless Governance for AI. Enforce policies in your ML training with one line of code.
# Venturalítica SDK ![coverage](https://img.shields.io/badge/coverage-100%25-brightgreen) **Frictionless Governance for AI Systems.** The Venturalítica SDK enables Data Scientists and ML Engineers to integrate compliance and risk management directly into their training workflows. Built on the **OSCAL** (Open Security Controls Assessment Language) standard, it provides semantic policy enforcement with educational audit trails. ## ✨ Key Features - **Glass Box Governance**: Sequential regulatory mapping (Art 9-15) for total transparency. - **Strict Mode**: Auto-enforcement of compliance checks in CI/CD environments. - **Deep Provenance**: Trace data lineage across Files, SQL, and S3 using `ArtifactProbe`. - **Local Sovereignty**: Zero-cloud dependency. All enforcement runs locally. - **TraceCollector Architecture**: Unified evidence gathering for BOM, metrics, anlogs. - **Educational Audits**: Control descriptions that explain *why* metrics matter. - **Deep Integrations**: Seamless "Glass Box" syncing with MLflow & WandB. - **OSCAL-Native**: Policy-as-Code using standard NIST formats. - **Annex IV Ready**: Auto-draft technical documentation from local traces. ## 📦 Installation ```bash pip install git+https://github.com/Venturalitica/venturalitica-sdk.git ``` ## ⚙️ Configuration The SDK supports the following Environment Variables. We recommend using a `.env` file (but **never commit it**!). | Variable | Description | Default | Required? | | :--- | :--- | :--- | :--- | | `MISTRAL_API_KEY` | [Get a Free Key](https://console.mistral.ai/). Used for Cloud Fallback if local Ollama fails. | None | **Recommended** | | `VENTURALITICA_LLM_PRO` | Set to `true` to use Mistral even if Ollama is available (Higher Quality). | `false` | No | | `VENTURALITICA_STRICT` | Set to `true` to enforce strict compliance checks (fail on missing metrics). | `false` | No | | `MLFLOW_TRACKING_URI` | If set, `monitor()` will auto-log audits to MLflow. | None | No | ## 📋 Prerequisites * **Python:** 3.11+ * **Local LLM (Optional):** * **Ollama**: (Recommended for standard local use). * **ALIA (Experimental)**: Native Spanish Sovereign model (Requires High-End GPU). * *Note: If you cannot run local models, please set `MISTRAL_API_KEY` for cloud generation.* ## 🚀 Quick Start ### 60-Second Demo ```python import venturalitica as vl # Auto-downloads UCI German Credit and runs bias audit results = vl.quickstart('loan') ``` **Output:** ``` [📊] Loaded: UCI Dataset #144 (1000 samples) [✅] PASSED: 3/3 fairness controls 🎉 Dataset passes bias checks! ``` ### Analyze Your Own Data First, create a **policy file** (`fairness.yaml`) that defines what to check: ```yaml assessment-plan: uuid: my-policy metadata: title: "Fairness Policy" reviewed-controls: control-selections: - include-controls: - control-id: gender-check description: "Approval rates must be similar across genders" props: - name: metric_key value: demographic_parity_diff - name: threshold value: "0.10" - name: operator value: "<" ``` Then run the audit: ```python import pandas as pd import venturalitica as vl df = pd.read_csv("my_data.csv") vl.enforce( data=df, target="approved", gender="gender", policy="fairness.yaml" ) ``` ## 📚 Documentation - **[Tutorial: Zero-Setup Audit](docs/tutorials/local-audit.md)**: "Hello World" - Scan & Visualize in 2 minutes - **[Tutorial: Training Integration](docs/training.md)**: Add compliance checks to your Python code - **[Concept: Strict Mode](docs/strict_mode.md)**: Enforcing compliance in CI/CD chains - **[Concept: The Regulatory Map](docs/compliance-dashboard.md)**: Understanding the Dashboard (Art 9-15) - **[Concept: Evidence Collection](docs/evidence-collection.md)**: How to record your audits - **[Samples Repository](https://github.com/venturalitica/venturalitica-sdk-samples)**: Real-world examples ## 🎯 Core Concepts ### Role-Based Binding The SDK uses a three-tier mapping system: 1. **Functional Roles** (defined by metrics): `target`, `prediction`, `dimension` 2. **Semantic Variables** (defined in policies): `gender`, `age_group`, `income` 3. **Physical Columns** (in your DataFrame): `sex_col`, `age_cat`, `salary` This decoupling allows policies to evolve independently of your training code. ### Educational Audits Control descriptions include regulatory context: ```yaml - control-id: data-quality-check description: "Data Quality: Minority class should represent at least 20% to avoid Class Imbalance" ``` ## 🛠️ CLI Tools ### BOM & Supply Chain The SDK automatically generates a **CycloneDX ML-BOM** during execution via `vl.monitor()`. **Detects:** - Python dependencies (`requirements.txt`, `pyproject.toml`) - ML models (scikit-learn, PyTorch, TensorFlow, XGBoost, etc.) - MLOps frameworks (MLflow, WandB, ClearML) **Output:** `bom` key within your audit trace JSON. ### Compliance Dashboard Launch the **Local Regulatory Map** to interpret your evidence: ```bash venturalitica ui ``` **[Read the Guide: Understanding the Dashboard](docs/compliance-dashboard.md)** **Features:** * **Article 9-15 Walk**: A sequential check of Risk, Data, Transparency, and Oversight. * **Sequential Verification**: See exactly which technical artifact satisfies which legal article. * **Annex IV Draft**: Generate the PDF-ready markdown file with `venturalitica doc`. **Integrates with:** * `trace_*.json` (from `vl.monitor()`) * `emissions.csv` (from CodeCarbon) * OSCAL policies ## 🔒 Data Sovereignty & Privacy Venturalítica follows a strict **Local-First** architecture. * **No Cloud Uploads**: `vl.enforce()` and `vl.quickstart()` run entirely on your local machine. Your datasets never leave your environment. * **Telemetry**: Usage metrics (if enabled) are strictly metadata (e.g., performance, error rates) and contain **NO PII**. * **Compliance Data**: All evidence (`trace_*.json`) is stored locally in `.venturalitica/`. You own your compliance data. ## ☁️ Venturalítica Cloud (Coming Soon) **Enterprise-grade EU AI Act & ISO 42001 compliance management** While the SDK provides frictionless local enforcement, **Venturalítica Cloud** will offer a complete compliance lifecycle management platform for **EU AI Act** and **ISO 42001**: ### What's Coming - **Visual Policy Builder**: Create OSCAL policies mapped to EU AI Act Articles 9-15 & ISO 42001 controls - **Team Collaboration**: Centralized policy management across organizations - **Compliance Dashboard**: Real-time status for EU AI Act & ISO 42001 requirements - **Annex IV Generator**: Auto-generate complete EU AI Act technical documentation - **Risk Assessment**: Guided workflows for high-risk AI system classification - **Audit Trail**: Immutable compliance history for regulatory inspections - **Integration Hub**: Connect with your existing MLOps and governance tools ### Early Access Interested in early access to Venturalítica Cloud? - **Join the waitlist**: [www.venturalitica.ai](http://www.venturalitica.ai) *(coming soon)* - **Enterprise inquiries**: [Contact us](http://www.venturalitica.ai) for pilot programs The SDK will always remain **free and open-source** under Apache 2.0. The cloud platform will offer additional enterprise features for teams managing EU AI Act and ISO 42001 compliance at scale. ## 🤝 Contributing We welcome contributions! Please see our [Contributing Guide](CONTRIBUTING.md). ## 📄 License Apache 2.0 - See [LICENSE](LICENSE) for details. ## 🔗 Links - [Samples Repository](https://github.com/venturalitica/venturalitica-sdk-samples) - [Documentation](docs/) - [OSCAL Standard](https://pages.nist.gov/OSCAL/)
text/markdown
null
Venturalítica <info@venturalitica.ai>
null
null
Apache-2.0
ai, compliance, eu-ai-act, fairness, governance, mlops, oscal
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Topic :: Scie...
[]
null
null
>=3.11
[]
[]
[]
[ "compliance-trestle>=3.11.0", "cyclonedx-python-lib>=11.6.0", "huggingface-hub", "numpy>=2.4.0", "pandas>=2.3.0", "posthog>=3.5.0", "pyyaml>=6.0.3", "rich>=14.2.0", "scikit-learn>=1.8.0", "typer>=0.21.0", "ucimlrepo>=0.0.7", "langchain-community; extra == \"agentic\"", "langchain-huggingface...
[]
[]
[]
[ "Homepage, https://venturalitica.ai", "Documentation, https://venturalitica.github.io/venturalitica-sdk", "Repository, https://github.com/Venturalitica/venturalitica-sdk", "Changelog, https://github.com/Venturalitica/venturalitica-sdk/releases" ]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:16:28.415022
venturalitica-0.5.0.tar.gz
114,913
ae/c5/64ba36b1e1efec485c8df1053d581f1946f829817df173e61a5b048874ef/venturalitica-0.5.0.tar.gz
source
sdist
null
false
45d30c556040fcf6767976cbe3371c9b
100d0c75924c14b3a929108330ff7d0913164315862c4842663d2d713f295de4
aec564ba36b1e1efec485c8df1053d581f1946f829817df173e61a5b048874ef
null
[ "LICENSE" ]
232
2.4
osint-security-threat-pkg
0.1.0
OSINT tool for security threats and global firepower.
# OSINT Security Threat Package A specialized Open Source Intelligence (OSINT) tool designed for cybersecurity threat intelligence and geopolitical military capability analysis. This package combines the power of the AlienVault Open Threat Exchange (OTX) API with an automated Global Firepower index scraper. ## Features | Module | Source | Description | |:---|:---|:---| | **OTX Scanner** | **AlienVault OTX** | Asynchronous wrapper for the OTX API. Fetches pulse counts, malware analyses, and threat intelligence for domains, IPs, and file hashes. | | **Firepower Scraper** | **Global Firepower** | Selenium-based automated scraper that extracts structured military statistics, financial metrics, manpower, and logistics data for any country. | ## Prerequisites * **Python 3.8+** * **Google Chrome:** Required for the `GlobalFirePowerScraper` (Selenium manages the ChromeDriver automatically via `webdriver-manager`). * **AlienVault API Key:** Required for the `OtxScanner`. ## Installation Navigate to the root directory of the package and install it via pip: ```bash cd OSINT-SECURITY-THREAT-PKG pip install osint-security-threat-pkg ``` This will automatically install necessary dependencies including selenium, beautifulsoup4, and the official OTXv2 Python SDK. ### Configuration To use the OTX Scanner, you must provide an AlienVault OTX API key. You can get a free key by signing up at otx.alienvault.com. Method 1: Environment Variable (Recommended) Set the variable in your terminal or .env file. The package will automatically detect it. Linux / macOS: ```bash export ALIENTVAULT_OTX_API_KEY="your_api_key_here" ``` Windows (PowerShell): ```PowerShell $env:ALIENTVAULT_OTX_API_KEY="your_api_key_here" ``` Method 2: Direct Initialization Pass the key directly when initializing the class in Python. ```Python from osint_security_threat_pkg import OtxScanner scanner = OtxScanner(api_key="your_api_key_here") ``` ### Usage Examples 1. Threat Intelligence (AlienVault OTX) Scan a domain asynchronously to check for associated malicious pulses and threat data. ```Python import asyncio from osint_security_threat_pkg import OtxScanner async def check_threats(): # Will use ALIENTVAULT_OTX_API_KEY from environment if not passed scanner = OtxScanner() domain = "pastebin.com" print(f"--- Scanning Domain: {domain} ---") result = await scanner.scan(domain) if result["success"]: data = result["data"] general_info = data.get("general", {}) pulse_info = general_info.get("pulse_info", {}) print(f"Total Pulses (Threat Reports): {pulse_info.get('count', 0)}") # Print the first few related pulses for pulse in pulse_info.get("pulses", [])[:3]: print(f"- {pulse.get('name')} (Tags: {', '.join(pulse.get('tags', []))})") else: print(f"Error: {result.get('error')}") if __name__ == "__main__": asyncio.run(check_threats()) ``` 2. Geopolitical Military Data (Global Firepower) Scrape comprehensive military strength data for a specific country. Using the with statement (context manager) ensures the headless browser closes automatically when finished. ```python from osint_security_threat_pkg import GlobalFirePowerScraper def check_military_strength(): country = "nigeria" print(f"--- Fetching Military Data for: {country.upper()} ---") # Initialize the scraper (headless=True runs the browser invisibly) with GlobalFirePowerScraper(headless=True) as scraper: # Search and optionally save the raw JSON output to a file data = scraper.search(country, save_to_file=False) if data: overview = data.get('overview', {}).get('ranks', {}) print(f"Global Firepower Rank: {overview.get('GFP Rank', 'N/A')}") manpower = data.get('manpower', {}).get('statistics', {}) if 'Available Manpower' in manpower: print(f"Available Manpower: {manpower['Available Manpower'].get('value')}") financials = data.get('financials', {}).get('statistics', {}) if 'Defense Budget' in financials: print(f"Defense Budget: {financials['Defense Budget'].get('value')}") else: print("Failed to retrieve data.") if __name__ == "__main__": check_military_strength() ``` ### Disclaimer This tool is intended for educational purposes and legitimate Open Source Intelligence (OSINT) research. The OtxScanner usage is subject to the AlienVault OTX Terms of Service. The GlobalFirePowerScraper automates interactions with a public website. Users are responsible for adhering to the site's robots.txt and avoiding excessive request rates that could degrade the service.
text/markdown
null
null
null
null
null
null
[]
[]
null
null
>=3.8
[]
[]
[]
[ "OTXv2>=1.5.0", "selenium>=4.10.0", "webdriver-manager>=4.0.0", "beautifulsoup4>=4.11.0" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.10.19
2026-02-19T12:16:26.861611
osint_security_threat_pkg-0.1.0.tar.gz
10,802
5c/d8/52f86fecfc66d5f420a485cd717cd48692e72cf369d8fa340422418cf866/osint_security_threat_pkg-0.1.0.tar.gz
source
sdist
null
false
1a7ba4dd4e460e605945a696fb53f7f4
521923803c5c07f119dfafe9d67d8a68b2066cd22adb846c56cbd6c71d31ed17
5cd852f86fecfc66d5f420a485cd717cd48692e72cf369d8fa340422418cf866
null
[]
246
2.1
odoo-addon-l10n-ro-stock-account
19.0.0.12.0
Romania - Stock Accounting
.. image:: https://odoo-community.org/readme-banner-image :target: https://odoo-community.org/get-involved?utm_source=readme :alt: Odoo Community Association ========================== Romania - Stock Accounting ========================== .. !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! !! This file is generated by oca-gen-addon-readme !! !! changes will be overwritten. !! !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! !! source digest: sha256:ae4c6eedd83892638efba1c1f2308e9dd23771a7830e9cac47ed232459a9f728 !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! .. |badge1| image:: https://img.shields.io/badge/maturity-Mature-brightgreen.png :target: https://odoo-community.org/page/development-status :alt: Mature .. |badge2| image:: https://img.shields.io/badge/license-AGPL--3-blue.png :target: http://www.gnu.org/licenses/agpl-3.0-standalone.html :alt: License: AGPL-3 .. |badge3| image:: https://img.shields.io/badge/github-OCA%2Fl10n--romania-lightgray.png?logo=github :target: https://github.com/OCA/l10n-romania/tree/19.0/l10n_ro_stock_account :alt: OCA/l10n-romania .. |badge4| image:: https://img.shields.io/badge/weblate-Translate%20me-F47D42.png :target: https://translation.odoo-community.org/projects/l10n-romania-19-0/l10n-romania-19-0-l10n_ro_stock_account :alt: Translate me on Weblate .. |badge5| image:: https://img.shields.io/badge/runboat-Try%20me-875A7B.png :target: https://runboat.odoo-community.org/builds?repo=OCA/l10n-romania&target_branch=19.0 :alt: Try me on Runboat |badge1| |badge2| |badge3| |badge4| |badge5| This module provides Romanian-specific stock accounting features that align with Romanian accounting standards and regulations. Below are the configuration options available. Overview -------- The module extends Odoo's standard stock accounting to meet Romanian accounting requirements, providing: - Location-specific accounting configurations - Romanian-specific stock valuation accounts - Product category stock account customization - Warehouse fiscal position management - Specialized accounts for stock operations **Table of contents** .. contents:: :local: Bug Tracker =========== Bugs are tracked on `GitHub Issues <https://github.com/OCA/l10n-romania/issues>`_. In case of trouble, please check there if your issue has already been reported. If you spotted it first, help us to smash it by providing a detailed and welcomed `feedback <https://github.com/OCA/l10n-romania/issues/new?body=module:%20l10n_ro_stock_account%0Aversion:%2019.0%0A%0A**Steps%20to%20reproduce**%0A-%20...%0A%0A**Current%20behavior**%0A%0A**Expected%20behavior**>`_. Do not contact contributors directly about support or help with technical issues. Credits ======= Authors ------- * NextERP Romania * Dorin Hongu * Forest and Biomass Romania Contributors ------------ - `NextERP Romania <https://www.nexterp.ro>`__: - Fekete Mihai <feketemihai@nexterp.ro> - Alexandru Teodor <teodoralexandru@nexterp.ro> - `Terrabit <https://www.terrabit.ro>`__: - Dorin Hongu <dhongu@gmail.com> Do not contact contributors directly about support or help with technical issues. Maintainers ----------- This module is maintained by the OCA. .. image:: https://odoo-community.org/logo.png :alt: Odoo Community Association :target: https://odoo-community.org OCA, or the Odoo Community Association, is a nonprofit organization whose mission is to support the collaborative development of Odoo features and promote its widespread use. .. |maintainer-dhongu| image:: https://github.com/dhongu.png?size=40px :target: https://github.com/dhongu :alt: dhongu .. |maintainer-feketemihai| image:: https://github.com/feketemihai.png?size=40px :target: https://github.com/feketemihai :alt: feketemihai Current `maintainers <https://odoo-community.org/page/maintainer-role>`__: |maintainer-dhongu| |maintainer-feketemihai| This module is part of the `OCA/l10n-romania <https://github.com/OCA/l10n-romania/tree/19.0/l10n_ro_stock_account>`_ project on GitHub. You are welcome to contribute. To learn how please visit https://odoo-community.org/page/Contribute.
text/x-rst
NextERP Romania,Dorin Hongu,Forest and Biomass Romania,Odoo Community Association (OCA)
support@odoo-community.org
null
null
AGPL-3
null
[ "Programming Language :: Python", "Framework :: Odoo", "Framework :: Odoo :: 19.0", "License :: OSI Approved :: GNU Affero General Public License v3", "Development Status :: 6 - Mature" ]
[]
https://github.com/OCA/l10n-romania
null
null
[]
[]
[]
[ "odoo-addon-l10n_ro_stock==19.0.*", "odoo==19.0.*" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.12.3
2026-02-19T12:15:47.181657
odoo_addon_l10n_ro_stock_account-19.0.0.12.0-py3-none-any.whl
72,786
43/de/6db1dd33be6e32e4faf77e55c8cf8a258e43cf556bca1331df634be00287/odoo_addon_l10n_ro_stock_account-19.0.0.12.0-py3-none-any.whl
py3
bdist_wheel
null
false
80c682c56e998e9e0b52cf42e7f8d103
e27248c844d8d077cfc943b66b83811ae181969f37c59c2ec94a1337a059c7cd
43de6db1dd33be6e32e4faf77e55c8cf8a258e43cf556bca1331df634be00287
null
[]
113
2.1
odoo-addon-l10n-ro-stock-account-date
19.0.0.2.0
Romania - Stock Accounting Date
.. image:: https://odoo-community.org/readme-banner-image :target: https://odoo-community.org/get-involved?utm_source=readme :alt: Odoo Community Association =============================== Romania - Stock Accounting Date =============================== .. !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! !! This file is generated by oca-gen-addon-readme !! !! changes will be overwritten. !! !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! !! source digest: sha256:f66dbbe262d11c281f964cf7ff3edab708783a511bfbbba3dd61d9c96c042e24 !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! .. |badge1| image:: https://img.shields.io/badge/maturity-Mature-brightgreen.png :target: https://odoo-community.org/page/development-status :alt: Mature .. |badge2| image:: https://img.shields.io/badge/license-AGPL--3-blue.png :target: http://www.gnu.org/licenses/agpl-3.0-standalone.html :alt: License: AGPL-3 .. |badge3| image:: https://img.shields.io/badge/github-OCA%2Fl10n--romania-lightgray.png?logo=github :target: https://github.com/OCA/l10n-romania/tree/19.0/l10n_ro_stock_account_date :alt: OCA/l10n-romania .. |badge4| image:: https://img.shields.io/badge/weblate-Translate%20me-F47D42.png :target: https://translation.odoo-community.org/projects/l10n-romania-19-0/l10n-romania-19-0-l10n_ro_stock_account_date :alt: Translate me on Weblate .. |badge5| image:: https://img.shields.io/badge/runboat-Try%20me-875A7B.png :target: https://runboat.odoo-community.org/builds?repo=OCA/l10n-romania&target_branch=19.0 :alt: Try me on Runboat |badge1| |badge2| |badge3| |badge4| |badge5| This module allows you to select "Accounting Date" for stock operations. Future dates are blocked for Romanian companies. You can restrict the dates from first day of previous month. **Table of contents** .. contents:: :local: Installation ============ To install this module, you need to: - clone the repository https://github.com/OCA/l10n-romania - add the path to this repository in your configuration (addons-path) - update the module list - search for "Romania - Stock Account Date" in your addons - install the module Bug Tracker =========== Bugs are tracked on `GitHub Issues <https://github.com/OCA/l10n-romania/issues>`_. In case of trouble, please check there if your issue has already been reported. If you spotted it first, help us to smash it by providing a detailed and welcomed `feedback <https://github.com/OCA/l10n-romania/issues/new?body=module:%20l10n_ro_stock_account_date%0Aversion:%2019.0%0A%0A**Steps%20to%20reproduce**%0A-%20...%0A%0A**Current%20behavior**%0A%0A**Expected%20behavior**>`_. Do not contact contributors directly about support or help with technical issues. Credits ======= Authors ------- * NextERP Romania Contributors ------------ - `NextERP Romania <https://www.nexterp.ro>`__: - Fekete Mihai <feketemihai@nexterp.ro> - Alexandru Teodor <teodoralexandru@nexterp.ro> - Cojocaru Marcel <cojocarumarcel@nexterp.ro> Do not contact contributors directly about support or help with technical issues. Maintainers ----------- This module is maintained by the OCA. .. image:: https://odoo-community.org/logo.png :alt: Odoo Community Association :target: https://odoo-community.org OCA, or the Odoo Community Association, is a nonprofit organization whose mission is to support the collaborative development of Odoo features and promote its widespread use. .. |maintainer-feketemihai| image:: https://github.com/feketemihai.png?size=40px :target: https://github.com/feketemihai :alt: feketemihai Current `maintainer <https://odoo-community.org/page/maintainer-role>`__: |maintainer-feketemihai| This module is part of the `OCA/l10n-romania <https://github.com/OCA/l10n-romania/tree/19.0/l10n_ro_stock_account_date>`_ project on GitHub. You are welcome to contribute. To learn how please visit https://odoo-community.org/page/Contribute.
text/x-rst
NextERP Romania,Odoo Community Association (OCA)
support@odoo-community.org
null
null
AGPL-3
null
[ "Programming Language :: Python", "Framework :: Odoo", "Framework :: Odoo :: 19.0", "License :: OSI Approved :: GNU Affero General Public License v3", "Development Status :: 6 - Mature" ]
[]
https://github.com/OCA/l10n-romania
null
null
[]
[]
[]
[ "odoo-addon-l10n_ro_stock_account==19.0.*", "odoo==19.0.*" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.12.3
2026-02-19T12:15:43.856932
odoo_addon_l10n_ro_stock_account_date-19.0.0.2.0-py3-none-any.whl
28,998
99/1d/7348c3d8ec5e919050d6aef534ccf801c0512e039c32cb73a187d27b83cd/odoo_addon_l10n_ro_stock_account_date-19.0.0.2.0-py3-none-any.whl
py3
bdist_wheel
null
false
279d2cd4dac354a7dbe5aced48972f05
3c351ece7c6e01e7fe308e65d0d4b1f9364e79dbe951ba4c884f3e972b3fa860
991d7348c3d8ec5e919050d6aef534ccf801c0512e039c32cb73a187d27b83cd
null
[]
99
2.4
workflows-mcp
9.2.0
MCP server for DAG-based workflow execution with YAML definitions and LLM collaboration
# Workflows MCP **Automate anything with simple YAML workflows for your AI assistant.** Workflows MCP is a [Model Context Protocol](https://modelcontextprotocol.io/) (MCP) server that lets you define powerful, reusable automation workflows in YAML and execute them through AI assistants like Claude. Think of it as GitHub Actions for your AI assistant—define your automation once, run it anywhere. --- ## Table of Contents - [What Does This Give Me?](#what-does-this-give-me) - [Why Should I Use This?](#why-should-i-use-this) - [Quick Start](#quick-start) - [How It Works](#how-it-works) - [What Can I Build?](#what-can-i-build) - [Creating Your First Workflow](#creating-your-first-workflow) - [Key Features](#key-features) - [Built-in Workflows](#built-in-workflows) - [Available MCP Tools](#available-mcp-tools) - [Configuration Reference](#configuration-reference) - [Examples](#examples) - [Development](#development) - [Troubleshooting](#troubleshooting) - [License](#license) --- ## What Does This Give Me? Workflows MCP transforms your AI assistant into an automation powerhouse. Instead of manually running commands or writing repetitive scripts, you define workflows in YAML, and your AI assistant executes them for you. **Real-world example:** ```text You: "Run the Python CI pipeline on my project" Claude: *Executes workflow that sets up environment, runs linting, and runs tests* Claude: "✓ All checks passed! Linting: ✓, Tests: ✓, Coverage: 92%" ``` --- ## Why Should I Use This? ### For Non-Technical Users - **No coding required** - Define automation in simple YAML - **Reusable templates** - Use pre-built workflows for common tasks - **AI-powered execution** - Just ask your AI assistant in plain English ### For Developers - **DRY principle** - Define once, use everywhere - **Parallel execution** - Automatic optimization of independent tasks - **Type-safe** - Validated inputs and outputs - **Composable** - Build complex workflows from simple building blocks ### For Teams - **Shared automation** - Version control your workflows - **Consistent processes** - Everyone uses the same tested workflows - **Custom templates** - Build company-specific automation libraries --- ## Quick Start ### Step 1: Install Choose one of these installation methods: **Option A: Using `uv` (Recommended - Faster)** ```bash uv pip install workflows-mcp ``` **Option B: Using `pip`** ```bash pip install workflows-mcp ``` **Requirements:** - Python 3.12 or higher - That's it! ### Step 2: Configure Your AI Assistant #### For Claude Desktop or Claude Code **Method 1: QTS Marketplace (Easiest)** Install the complete workflows plugin with agent, skills, and auto-configuration: ```bash # Add the marketplace (one-time setup) /plugin marketplace add qtsone/marketplace # Install the workflows plugin /plugin install workflows@qtsone ``` This automatically configures the MCP server to look for custom workflows in `~/.workflows` and `./.workflows` (optional directories you can create if you want to add custom workflows). **Method 2: Manual Configuration** Add this to your Claude Desktop config file: - **macOS**: `~/Library/Application Support/Claude/claude_desktop_config.json` - **Windows**: `%APPDATA%\Claude\claude_desktop_config.json` ```json { "mcpServers": { "workflows": { "command": "uvx", "args": ["workflows-mcp", "--refresh"], "env": { "WORKFLOWS_TEMPLATE_PATHS": "~/.workflows,./.workflows" } } } } ``` **Note:** The `WORKFLOWS_TEMPLATE_PATHS` directories are optional. Only create them if you want to add custom workflows. The server works perfectly fine with just the built-in workflows. We recommend using `~/.workflows` as it's also the default location for optional LLM configuration (`llm-config.yml`). #### For Other MCP-Compatible AI Assistants The configuration is similar. For example, Gemini CLI users would add this to `~/.gemini/settings.json` with the same structure. ### Step 3: Restart and Test 1. Restart your AI assistant (e.g., Claude Desktop) 2. Try it out: ```text You: "List all available workflows" ``` Your AI assistant will show you all the built-in workflows ready to use! --- ## How It Works Workflows MCP operates on three simple concepts: ### 1. **Workflows** (The What) YAML files that define what you want to automate. Each workflow has: - **Inputs** - Parameters users can customize - **Blocks** - Individual tasks to execute - **Outputs** - Results returned to the user ### 2. **Blocks** (The How) Individual tasks within a workflow. Available block types: - `Shell` - Run shell commands - `LLMCall` - Call AI/LLM APIs - `HttpCall` - Make HTTP requests - `CreateFile`, `ReadFiles`, `EditFile` - File operations - `Workflow` - Call other workflows (composition) - `Prompt` - Interactive user prompts - `ImageGen` - Generate and edit images (DALL-E, Stable Diffusion) - `ReadJSONState`, `WriteJSONState`, `MergeJSONState` - State management ### 3. **Execution** (The Magic) The server automatically: - Analyzes dependencies between blocks - Runs independent blocks in parallel - Handles errors gracefully - Substitutes variables dynamically **Example Flow:** ```yaml # Python CI Pipeline setup_env → run_linting ↘ → validate_results setup_env → run_tests ↗ ``` Tasks run in parallel when possible, saving time! --- ## What Can I Build? The possibilities are endless. Here are some examples: ### Development Automation - **CI/CD Pipelines** - Automated testing, linting, building - **Code Quality Checks** - Run multiple linters and formatters in parallel - **Deployment Workflows** - Build, test, and deploy applications ### Git Operations - **Smart Branch Management** - Create feature branches with proper naming - **Automated Commits** - Stage files and commit with generated messages - **Repository Analysis** - Analyze changes, detect patterns ### Data Processing - **File Transformations** - Process and transform files in batch - **API Orchestration** - Chain multiple API calls together - **Report Generation** - Generate reports from templates ### AI-Powered Tasks - **Content Analysis** - Use LLMs to analyze and extract insights - **Code Generation** - Generate code based on specifications - **Automated Review** - Review code, documents, or data --- ## Creating Your First Workflow Let's create a simple workflow that greets a user: ### 1. Create a YAML file First, create the workflows directory (if it doesn't exist): ```bash mkdir -p ~/.workflows ``` Then save this as `~/.workflows/greet-user.yaml`: ```yaml name: greet-user description: A friendly greeting workflow tags: [example, greeting] inputs: name: type: str description: Name of the person to greet default: "World" language: type: str description: Language for greeting (en, es, fr) default: "en" blocks: - id: create_greeting type: Shell inputs: command: | case "{{inputs.language}}" in es) echo "¡Hola, {{inputs.name}}!" ;; fr) echo "Bonjour, {{inputs.name}}!" ;; *) echo "Hello, {{inputs.name}}!" ;; esac outputs: greeting: value: "{{blocks.create_greeting.outputs.stdout}}" type: str description: The personalized greeting ``` ### 2. Restart your AI assistant The workflow is automatically discovered from `~/.workflows/` ### 3. Use it! ```text You: "Run the greet-user workflow with name=Alice and language=es" Claude: *Executes workflow* Claude: "¡Hola, Alice!" ``` ### Understanding the Workflow - **inputs** - Define customizable parameters - **blocks** - Each block is a task (here, running a shell command) - **{{inputs.name}}** - Variable substitution (replaced at runtime) - **outputs** - What gets returned to the user --- ## Key Features ### 🚀 Smart Parallel Execution The server automatically detects which tasks can run in parallel. Use `depends_on` to specify dependencies—independent tasks run concurrently for maximum efficiency. **See examples:** `tests/workflows/core/dag-execution/parallel-execution.yaml` ### 🔐 Secure Secrets Management Store sensitive data like API keys securely using environment variables. Secrets are resolved server-side and never exposed to the LLM context. ```json { "mcpServers": { "workflows": { "env": { "WORKFLOW_SECRET_GITHUB_TOKEN": "ghp_your_token_here", "WORKFLOW_SECRET_OPENAI_API_KEY": "sk-your_key_here" } } } } ``` Use in workflows: `{{secrets.GITHUB_TOKEN}}` **Security features:** - ✅ Secrets never appear in LLM context - ✅ Automatic redaction in all outputs - ✅ Server-side resolution only - ✅ Fail-fast on missing secrets **See examples:** `tests/workflows/core/secrets/` ### 🎨 Full Jinja2 Template Support All workflow fields are Jinja2 templates with support for: - **Variable expressions**: `{{inputs.name}}`, `{{blocks.test.outputs.result}}` - **Control structures**: `{% if condition %}...{% endif %}`, `{% for item in list %}...{% endfor %}` - **Custom filters**: `quote`, `prettyjson`, `b64encode`, `hash`, `trim`, `upper`, `lower`, `replace` - **Global functions**: `now()`, `render()`, `get()`, `len()`, `range()` - **Safe accessor**: `get(obj, 'path.to.key', default)` - dotted paths, JSON auto-parse, never throws - **Filter chaining**: `{{inputs.text | trim | lower | replace(' ', '_')}}` **See examples:** `tests/workflows/core/filters/filters-chaining.yaml` ### 📂 ReadFiles Block with Outline Extraction Read files with glob patterns, multiple output modes, and automatic outline extraction: **Features:** - Glob pattern support (`*.py`, `**/*.ts`) - Three output modes: - `full` - Complete file content - `outline` - Structural outline (90-97% context reduction for Python/Markdown) - `summary` - Outline + docstrings/comments - Gitignore integration and file filtering - Size limits and file count limits - Multi-file reading in single block - Smart output format: - Single file: Direct content (string) - Multiple files: YAML-formatted structure - No files: Empty string **See examples:** `tests/workflows/core/file-operations/readfiles-test.yaml` ### ✏️ Deterministic File Editing The `EditFile` block provides powerful deterministic file editing with multiple operation strategies: **Features:** - 6 operation types (replace_text, replace_lines, insert_lines, delete_lines, patch, regex_replace) - Atomic transactions (all-or-nothing by default) - Automatic backup creation before editing - Dry-run mode for previewing changes - Comprehensive diff generation - Path traversal protection **See examples:** `tests/workflows/core/file-operations/editfile-operations-test.yaml` ### 🤖 LLM Integration Call AI models directly from workflows with automatic retry and validation. Configure providers using `~/.workflows/llm-config.yml` (optional). **Supported providers:** OpenAI, Anthropic, Gemini, Ollama, OpenAI-compatible (LM Studio, vLLM) **Example:** ```yaml - id: analyze type: LLMCall inputs: profile: default # Existing profile from ~/.workflows/llm-config.yml prompt: "Analyze this text: {{inputs.text}}" response_schema: type: object required: [sentiment, summary] properties: sentiment: type: string enum: [positive, negative, neutral] summary: type: string ``` **Profile Fallback for Portable Workflows:** When a workflow requests a profile that doesn't exist in your config, the system automatically falls back to `default_profile` with a warning. This enables **workflow portability** - authors can write workflows with semantic profile names (like `cloud-mini`, `cloud-thinking`, `local`) without requiring specific user configurations. ```yaml # ~/.workflows/llm-config.yml profiles: my-model: provider: openai-cloud model: gpt-4o max_tokens: 4000 default_profile: my-model ``` ### 🎨 Image Generation Generate, edit, and create variations of images using OpenAI DALL-E or compatible providers (like local Stable Diffusion via OpenAI-compatible API). **Key Features:** - **Model Capability System**: Automatic validation of operations and parameters based on model support - **Profile Support**: Use `~/.workflows/llm-config.yml` to manage providers and models (same as LLMCall) - **Direct File Saving**: Save generated images directly to disk with `output_file` - **Pluggable Providers**: Support for OpenAI (DALL-E 2/3) and any OpenAI-compatible image API **Model Compatibility:** The executor validates operations and parameters based on model capabilities: | Model | Generate | Edit | Variation | Optional Parameters | |-------|----------|------|-----------|---------------------| | `dall-e-3` | ✓ | ✗ | ✗ | `response_format`, `quality`, `style` | | `dall-e-2` | ✓ | ✓ | ✓ | `response_format` | | `gpt-image-1` | ✓ | ✓ | ✗ | None | | `gpt-image-*` | ✓ | ✓ | ✗ | None | The executor automatically filters parameters and provides clear error messages when operations are not supported. **Configuration:** | Input | Description | Default | |-------|-------------|---------| | `prompt` | Text description (required for generate/edit) | - | | `profile` | Profile name from config (recommended) | `default` | | `operation` | `generate`, `edit`, or `variation` | `generate` | | `model` | Model name (e.g., `dall-e-3`) | `dall-e-3` | | `size` | Image dimensions (e.g., `1024x1024`) | `1024x1024` | | `quality` | `standard` or `hd` (dall-e-3 only) | `standard` | | `style` | `vivid` or `natural` (dall-e-3 only) | `vivid` | | `response_format` | `url` or `b64_json` | `url` | | `n` | Number of images to generate | `1` | | `output_file` | Path to save image (e.g., `{{tmp}}/img.png`) | - | | `image` | Path to base image (for edit/variation) | - | | `mask` | Path to mask image (transparent areas define where to edit) | - | **Examples:** **1. Basic Generation (using profile):** ```yaml - id: generate_art type: ImageGen inputs: prompt: "A cyberpunk city at night" profile: default size: "1024x1024" output_file: "{{tmp}}/city.png" ``` **2. Image Editing with DALL-E 2:** ```yaml - id: edit_photo type: ImageGen inputs: operation: edit model: dall-e-2 # DALL-E 3 does not support edit prompt: "Add a red hat to the person" image: "/path/to/photo.png" mask: "/path/to/mask.png" output_file: "{{tmp}}/edited.png" ``` **3. Using Custom Provider (e.g., Local SD):** ```yaml - id: local_gen type: ImageGen inputs: prompt: "A medieval castle" provider: openai_compatible api_url: "http://localhost:8000/v1" model: "sd-xl" output_file: "{{tmp}}/castle.png" ``` ### 🔁 Universal Iteration (for_each) Iterate over collections with ANY block type using `for_each`. Supports parallel and sequential execution modes with error handling. **Iteration variables:** - `{{each.key}}` - Current key - `{{each.value}}` - Current value - `{{each.index}}` - Zero-based position - `{{each.count}}` - Total iterations **See examples:** `tests/workflows/core/for_each/for-each-comprehensive.yaml` ### 🔄 Workflow Composition Build complex workflows from simple reusable pieces using the `Workflow` block type. Supports recursion with configurable depth limits. **See examples:** `tests/workflows/core/composition/` ### 📝 Conditional Execution Run blocks only when conditions are met using the `condition` field. Conditions are evaluated as Jinja2 expressions. **See examples:** `tests/workflows/core/conditionals/` ### 💬 Interactive Workflows Pause workflows to get user input using the `Prompt` block. Use `resume_workflow(job_id, response)` to continue execution with the user's input. **See examples:** `tests/workflows/interactive-simple-approval.yaml` ### ⚡ Async Execution Execute long-running workflows without blocking using `mode="async"`. Track progress with `get_job_status(job_id)` and cancel with `cancel_job(job_id)`. **Use cases:** - Long CI/CD pipelines - Large-scale data processing - Multi-stage deployments - Resource-intensive analysis --- ## Built-in Workflows The server includes many ready-to-use workflows for common tasks. ### 📋 Quality & Testing **Core workflows** (Python CI, Git operations, file operations) are actively used and thoroughly tested. Some advanced workflows are still being refined. **Best practice:** Always inspect a workflow before using it (`get_workflow_info` tool) and test on non-production systems first. **Battle-tested examples:** The workflows in `tests/workflows/core/` are comprehensively tested in CI and demonstrate all core features reliably. ### Discovering Workflows **List all workflows:** ```text You: "List all available workflows" ``` **Get detailed information:** ```text You: "Show me details about the python-ci-pipeline workflow" ``` **Filter by category:** ```text You: "List workflows tagged with 'python'" You: "Show me all git workflows" ``` **Popular workflows include:** Python CI pipelines, Git operations (checkout, commit, status), linting tools, test runners, and file operations. --- ## Available MCP Tools When you configure workflows-mcp, your AI assistant gets these tools: ### Workflow Execution - **execute_workflow** - Run a registered workflow by name - `mode`: "sync" (default) or "async" - `timeout`: Optional timeout in seconds for async mode (default: 3600, max: 86400) - **execute_inline_workflow** - Execute YAML directly without registration ### Workflow Discovery - **list_workflows** - List all available workflows (optional tag filtering) - **get_workflow_info** - Get detailed information about a workflow ### Workflow Validation - **validate_workflow_yaml** - Validate YAML before execution - **get_workflow_schema** - Get the complete JSON schema ### Job Management - **get_job_status** - Get status and outputs of a workflow job - **cancel_job** - Cancel a pending or running job - **list_jobs** - List workflow jobs with optional filtering - **get_queue_stats** - Get queue statistics for monitoring - **resume_workflow** - Resume a paused workflow --- ## Configuration Reference ### Environment Variables Configure the server behavior with these environment variables: | Variable | Description | Default | Range | |----------|-------------|---------|-------| | `WORKFLOWS_TEMPLATE_PATHS` | Comma-separated workflow directories | *(none)* | Valid paths | | `WORKFLOWS_MAX_RECURSION_DEPTH` | Maximum workflow recursion depth | `50` | `1-10000` | | `WORKFLOWS_LOG_LEVEL` | Logging verbosity | `INFO` | DEBUG, INFO, WARNING, ERROR, CRITICAL | | `WORKFLOW_SECRET_<NAME>` | Secret value (e.g., `WORKFLOW_SECRET_API_KEY`) | *(none)* | Any string | | `WORKFLOWS_IO_QUEUE_ENABLED` | Enable serialized I/O operations | `true` | true, false | | `WORKFLOWS_JOB_QUEUE_ENABLED` | Enable async workflow execution | `true` | true, false | | `WORKFLOWS_JOB_QUEUE_WORKERS` | Worker pool size for async jobs | `3` | 1-100 | | `WORKFLOWS_MAX_CONCURRENT_JOBS` | Maximum concurrent jobs | `100` | 1-10000 | ### Example Configuration ```json { "mcpServers": { "workflows": { "command": "uvx", "args": ["workflows-mcp", "--refresh"], "env": { "WORKFLOWS_TEMPLATE_PATHS": "~/.workflows,./project-workflows", "WORKFLOWS_LOG_LEVEL": "DEBUG", "WORKFLOWS_MAX_RECURSION_DEPTH": "100", "WORKFLOWS_IO_QUEUE_ENABLED": "true", "WORKFLOWS_JOB_QUEUE_ENABLED": "true", "WORKFLOWS_JOB_QUEUE_WORKERS": "5", "WORKFLOWS_MAX_CONCURRENT_JOBS": "200", "WORKFLOW_SECRET_GITHUB_TOKEN": "ghp_xxxxx", "WORKFLOW_SECRET_OPENAI_API_KEY": "sk-xxxxx" } } } } ``` ### Custom Workflow Directories The server loads workflows from: 1. Built-in templates (always loaded) 2. Custom directories (specified in `WORKFLOWS_TEMPLATE_PATHS`, optional) **Note:** Custom workflow directories are not created automatically. You need to create them manually if you want to add your own workflows. The server works fine without them using only built-in workflows. **Load order priority:** Later directories override earlier ones by workflow name. --- ## Examples For comprehensive examples demonstrating all features, see the test workflows in `tests/workflows/core/`: - **File operations**: `tests/workflows/core/file-operations/` - **Parallel execution**: `tests/workflows/core/dag-execution/` - **Conditionals**: `tests/workflows/core/conditionals/` - **Composition**: `tests/workflows/core/composition/` - **Secrets**: `tests/workflows/core/secrets/` - **Filters**: `tests/workflows/core/filters/` - **Iteration**: `tests/workflows/core/for_each/` ### Quick Example: Simple Shell Command ```yaml name: disk-usage description: Check disk usage tags: [utility, system] blocks: - id: check_disk type: Shell inputs: command: "df -h" outputs: disk_info: value: "{{blocks.check_disk.outputs.stdout}}" type: str ``` --- ## Development ### Running Tests ```bash # Install development dependencies uv sync --all-extras # Run all tests uv run pytest # With coverage uv run pytest --cov=workflows_mcp --cov-report=term-missing # Run specific test uv run pytest tests/test_mcp_client.py -v ``` ### Code Quality ```bash # Type checking uv run mypy src/workflows_mcp/ # Linting uv run ruff check src/workflows_mcp/ # Formatting uv run ruff format src/workflows_mcp/ ``` ### Testing the MCP Server For interactive testing and debugging: **1. Create `.mcp.json`:** ```json { "mcpServers": { "workflows": { "command": "uv", "args": ["run", "workflows-mcp"], "env": { "WORKFLOWS_LOG_LEVEL": "DEBUG", "WORKFLOWS_TEMPLATE_PATHS": "~/.workflows" } } } } ``` **2. Run MCP Inspector:** ```bash npx @modelcontextprotocol/inspector --config .mcp.json --server workflows ``` This opens a web interface for testing tool calls and debugging workflow execution. ### Project Structure ```bash workflows-mcp/ ├── src/workflows_mcp/ # Main source code │ ├── engine/ # Workflow execution engine │ │ ├── executor_base.py # Base executor class │ │ ├── executors_core.py # Shell, Workflow executors │ │ ├── executors_file.py # File operation executors (CreateFile, ReadFiles, EditFile) │ │ ├── file_outline.py # File outline extraction utilities │ │ ├── executors_http.py # HTTP call executor │ │ ├── executors_llm.py # LLM call executor │ │ ├── executors_state.py # State management executors │ │ ├── workflow_runner.py # Main workflow orchestrator │ │ ├── dag.py # DAG resolution │ │ ├── resolver/ # Unified variable resolver (Jinja2) │ │ └── secrets/ # Secrets management │ ├── templates/ # Built-in workflow templates │ │ ├── python/ # Python workflows │ │ ├── git/ # Git workflows │ │ ├── node/ # Node.js workflows │ │ └── ... │ ├── server.py # MCP server setup │ ├── tools.py # MCP tool implementations │ └── __main__.py # Entry point ├── tests/ # Test suite ├── pyproject.toml # Project configuration └── README.md # This file ``` --- ## Troubleshooting ### Installation Issues **Problem:** `command not found: workflows-mcp` **Solution:** ```bash # Ensure Python 3.12+ is installed python --version # Should be 3.12 or higher # Reinstall with uv uv pip install --force-reinstall workflows-mcp # Or verify installation pip show workflows-mcp ``` **Problem:** Python version too old **Solution:** ```bash # Install Python 3.12+ using your package manager # macOS (Homebrew) brew install python@3.12 # Ubuntu/Debian sudo apt install python3.12 # Update uv to use Python 3.12 uv venv --python 3.12 ``` ### Configuration Issues **Problem:** Workflows not loading in Claude **Solution:** 1. Verify config file location (see [Quick Start](#quick-start)) 2. Check JSON syntax with a validator 3. Restart Claude Desktop completely 4. Check Claude Desktop logs: - macOS: `~/Library/Logs/Claude/` - Windows: `%APPDATA%\Claude\logs\` **Problem:** Custom workflows not found **Solution:** ```bash # First, make sure you created the directory mkdir -p ~/.workflows # Verify WORKFLOWS_TEMPLATE_PATHS is correct # Paths should exist and contain .yaml files # Check directory exists and contains workflows ls ~/.workflows/ # Check YAML syntax python -c "import yaml; yaml.safe_load(open('~/.workflows/my-workflow.yaml'))" ``` ### Workflow Execution Issues **Problem:** Workflow fails with "not found" error **Solution:** ```text You: "List all workflows" # This shows exact workflow names # Use the exact name from the list ``` **Problem:** Variables not substituting **Solution:** - Check syntax: `{{inputs.name}}` not `{inputs.name}` - Ensure input is defined in `inputs:` section - For block outputs: `{{blocks.block_id.outputs.field_name}}` **Problem:** Secrets not working **Solution:** 1. Check environment variable name: `WORKFLOW_SECRET_<NAME>` 2. Reference in workflow: `{{secrets.NAME}}` (without prefix) 3. Verify secrets are in MCP server config, not workflow YAML 4. Restart MCP server after adding secrets ### Performance Issues **Problem:** Workflows running slowly **Solution:** - Check if tasks can run in parallel (remove unnecessary `depends_on`) - Enable debug logging to see execution waves: ```text You: "Run workflow X with debug=true" ``` - Review task dependencies—too many serialized tasks slow execution **Problem:** Shell commands timing out **Solution:** ```yaml blocks: - id: long_task type: Shell inputs: command: "./long-script.sh" timeout: 600 # Increase timeout (default: 120 seconds) ``` ### Debug Mode Enable detailed logging for troubleshooting: **Method 1: Environment variable** ```json { "mcpServers": { "workflows": { "env": { "WORKFLOWS_LOG_LEVEL": "DEBUG" } } } } ``` **Method 2: Per-execution debug** ```text You: "Run python-ci-pipeline with debug=true" ``` Debug logs are written to `/tmp/<workflow>-<timestamp>.json` with: - Block execution details - Variable resolution steps - DAG wave analysis - Timing information --- ## Architecture Workflows MCP uses a **fractal execution model** where workflows and blocks share the same execution context structure. This enables clean composition and recursive workflows. ### Key Components - **WorkflowRunner** - Orchestrates workflow execution - **BlockOrchestrator** - Executes individual blocks with error handling - **DAGResolver** - Resolves dependencies and computes parallel execution waves - **UnifiedVariableResolver** - Jinja2-based variable resolution with four namespaces: - `inputs` - Workflow runtime inputs - `blocks` - Block outputs and metadata - `metadata` - Workflow metadata - `secrets` - Server-side secrets (never exposed to LLM) ### Execution Model Workflows execute in **waves**—groups of blocks that can run in parallel: ```text Wave 1: [setup] Wave 2: [lint, test] ← Parallel execution Wave 3: [validate] ``` This maximizes efficiency by running independent tasks concurrently. --- ## Contributing We welcome contributions! Here's how you can help: ### Report Issues - [GitHub Issues](https://github.com/qtsone/workflows-mcp/issues) ### Contribute Workflows 1. Create a new workflow in appropriate category 2. Test thoroughly 3. Submit a pull request ### Improve Documentation - Fix typos or unclear explanations - Add examples - Improve troubleshooting guides ### Code Contributions - Follow existing code style - Add tests for new features - Update documentation --- ## Links - **GitHub**: [github.com/qtsone/workflows-mcp](https://github.com/qtsone/workflows-mcp) - **Issues**: [github.com/qtsone/workflows-mcp/issues](https://github.com/qtsone/workflows-mcp/issues) - **Changelog**: [CHANGELOG.md](./CHANGELOG.md) - **MCP Protocol**: [modelcontextprotocol.io](https://modelcontextprotocol.io/) --- ## License [AGPL-3.0-or-later](./LICENSE) ### Workflow Files and Your IP **YAML workflow files that use the workflows-mcp engine are NOT considered derivative works under AGPL-3.0.** Users retain full ownership and may license their workflow files under any terms they choose. This clarification applies to: - `.yaml`/`.yml` workflow definition files - Configuration files (e.g., `llm-config.yml`) - Workflow documentation and examples you create This does NOT apply to: - Modifications to the engine source code - Custom block executors (Python code) - Forks of the engine itself --- ## FAQ ### Do I need to know Python to use this? No! You only need to: 1. Install the package (one command) 2. Configure your AI assistant (copy-paste JSON) 3. Write simple YAML workflows (or use built-in ones) ### Do I need to know what MCP is? No! Just think of it as a way for your AI assistant to run workflows. The technical details are handled for you. ### Can I use this without Claude? Yes! Any MCP-compatible AI assistant can use workflows-mcp. The configuration is similar across different assistants. ### Are workflows secure? Yes! The server includes: - Server-side secret resolution (secrets never reach the AI) - Automatic redaction of sensitive data - Sandboxed execution contexts - Audit logging ### Are all built-in workflows production-ready? The core workflows (Python CI, Git operations, basic file operations) are actively used and reliable. Some advanced workflows are still being refined and tested. **Best practice:** Always inspect a workflow before using it (`get_workflow_info` tool) and test on non-production systems first. The workflows in `tests/workflows/` are thoroughly tested in CI and are great examples to learn from. ### Can I share workflows with my team? Absolutely! Workflows are just YAML files. You can: - Commit them to version control - Share them in a company repository - Publish them as packages ### What's the performance like? Excellent! The DAG-based execution model automatically parallelizes independent tasks. Many users see 2-3x speedup compared to sequential execution. ### Can workflows call other workflows? Yes! Use the `Workflow` block type to compose workflows. Recursion is supported with configurable depth limits. ### How do I get help? 1. Check [Troubleshooting](#troubleshooting) 2. Search [GitHub Issues](https://github.com/qtsone/workflows-mcp/issues) 3. Open a new issue with details --- **Ready to automate?** Install workflows-mcp and start building powerful automation workflows today! 🚀
text/markdown
null
Iulian Bacalu <ibacalu@icloud.com>
null
null
null
automation, claude, dag, llm-collaboration, mcp, workflow
[ "Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: GNU Affero General Public License v3 or later (AGPLv3+)", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.12", "Topic :: Software Development :: Libraries :: Python Mod...
[]
null
null
>=3.12
[]
[]
[]
[ "cryptography>=41.0.0", "jinja2>=3.1.0", "jsonschema>=4.0", "mcp[cli]>=1.26.0", "openai>=1.68.0", "pathspec>=0.12.0", "pydantic>=2.0", "pyjwt[crypto]>=2.8.0", "pyyaml>=6.0", "sqlite-vec>=0.1.6", "tree-sitter-languages>=1.10.0", "types-pyyaml>=6.0.12.20250915", "mypy>=1.8; extra == \"dev\"", ...
[]
[]
[]
[ "Homepage, https://github.com/qtsone/workflows-mcp", "Repository, https://github.com/qtsone/workflows-mcp", "Documentation, https://github.com/qtsone/workflows-mcp/blob/main/README.md", "Issues, https://github.com/qtsone/workflows-mcp/issues", "Changelog, https://github.com/qtsone/workflows-mcp/blob/main/CH...
twine/6.1.0 CPython/3.13.7
2026-02-19T12:14:37.239720
workflows_mcp-9.2.0.tar.gz
425,715
4a/76/1cc28454708ebbf87421650de6e40409749e74686d073974c0b945589f06/workflows_mcp-9.2.0.tar.gz
source
sdist
null
false
adb4548af0f0a2c2fefceb860c93b010
a3c14184901d85521b54692a7bf76c60b7b10b712f208d298a1e6d2a1c681181
4a761cc28454708ebbf87421650de6e40409749e74686d073974c0b945589f06
AGPL-3.0-or-later
[ "LICENSE" ]
234
2.4
pymodelserve
0.1.1
Run ML models in isolated subprocess environments with automatic dependency management
# pymodelserve Run ML models in isolated subprocess environments with automatic dependency management. Define your model once, and the library handles virtual environment creation, dependency installation, and inter-process communication via named pipes. Supports TensorFlow, PyTorch, or any Python ML framework with optional Django integration. ## Installation ```bash pip install pymodelserve ``` For Django integration: ```bash pip install pymodelserve[django] ``` ## Quick Start ### 1. Create a Model Create a directory structure for your model: ``` models/my_classifier/ ├── model.yaml # Configuration ├── model.py # Client implementation ├── requirements.txt # Dependencies └── weights/ # Model files (optional) ``` **model.yaml:** ```yaml name: my_classifier version: "1.0.0" python: ">=3.11" client: module: model class: MyClassifierClient requirements: requirements.txt handlers: - name: classify input: image_path: string output: class: string confidence: float ``` **model.py:** ```python from pymodelserve import ModelClient, handler class MyClassifierClient(ModelClient): def setup(self): # Load your model here import tensorflow as tf self.model = tf.keras.models.load_model("weights/model.keras") @handler("classify") def classify(self, image_path: str) -> dict: # Your classification logic return {"class": "cat", "confidence": 0.95} if __name__ == "__main__": MyClassifierClient().run() ``` ### 2. Use the Model ```python from pymodelserve import ModelManager # Using context manager (recommended) with ModelManager.from_yaml("./models/my_classifier/model.yaml") as model: result = model.request("classify", {"image_path": "/path/to/image.jpg"}) print(result) # {"class": "cat", "confidence": 0.95} # Or manual lifecycle manager = ModelManager.from_yaml("./models/my_classifier/model.yaml") manager.start() result = manager.request("classify", {"image_path": "/path/to/image.jpg"}) manager.stop() ``` ### 3. Serve Models (CLI) ```bash # Serve a single model pml serve ./models/my_classifier/ # Serve all models in a directory pml serve ./models/ --all # List discovered models pml list ./models/ # Create a new model scaffold pml init my_new_model --framework tensorflow ``` ## Features - **Isolated Environments**: Each model runs in its own virtual environment with isolated dependencies - **Named Pipe IPC**: Fast inter-process communication with no network overhead - **Auto-Discovery**: Scan directories for models and register them automatically - **Health Monitoring**: Periodic health checks with automatic restart on failure - **Django Integration**: Optional Django app with views and management commands - **CLI Tools**: Commands for serving, testing, and managing models ## Model Registry Manage multiple models with the registry: ```python from pymodelserve import ModelRegistry registry = ModelRegistry() registry.register("fruit", "./models/fruit_classifier/") registry.register("sentiment", "./models/sentiment/") # Start all models registry.start_all() # Use specific model result = registry.get("fruit").request("classify", {"image_path": "..."}) # Get status print(registry.status()) # Stop all registry.stop_all() ``` Or use auto-discovery: ```python from pymodelserve import discover_models, ModelRegistry # Discover all models configs = discover_models("./models/") # Create registry from discovered models registry = ModelRegistry() registry.register_from_dir("./models/") ``` ## Health Monitoring ```python from pymodelserve import ModelRegistry from pymodelserve.health import HealthChecker registry = ModelRegistry() registry.register_from_dir("./models/") registry.start_all() # Start health monitoring with auto-restart checker = HealthChecker( registry=registry, interval=30, # Check every 30 seconds max_failures=3, # Restart after 3 failures auto_restart=True, ) checker.start() # ... your application runs ... checker.stop() registry.stop_all() ``` ## Django Integration **settings.py:** ```python INSTALLED_APPS = [ ... 'pymodelserve.contrib.django', ] MLSERVE = { "models_dir": BASE_DIR / "ml_models", "auto_start": True, "health_check_interval": 30, } ``` **views.py:** ```python from pymodelserve.contrib.django.views import ModelAPIView class ClassifyImageView(ModelAPIView): model_name = "fruit_classifier" handler = "classify" def get_handler_input(self, request): image = request.FILES["image"] path = save_uploaded_file(image) return {"image_path": str(path)} ``` **urls.py:** ```python from django.urls import path from pymodelserve.contrib.django.views import GenericModelView, ModelStatusView urlpatterns = [ # Generic endpoint for any model/handler path("api/models/<str:model_name>/<str:handler>/", GenericModelView.as_view()), # Status endpoint path("api/models/status/", ModelStatusView.as_view()), ] ``` **Management command:** ```bash python manage.py serve_models python manage.py serve_models --model fruit_classifier --model sentiment ``` ## Client Implementation The `ModelClient` base class runs in the subprocess and handles IPC: ```python from pymodelserve import ModelClient, handler class MyModelClient(ModelClient): def setup(self): """Called once after IPC is established, before processing requests.""" self.model = load_my_model() def teardown(self): """Called when shutting down.""" cleanup_resources() @handler("predict") def predict(self, x: float, y: float) -> dict: """Handler methods receive kwargs and return dicts.""" result = self.model.predict([[x, y]]) return {"prediction": float(result[0])} # Alternative: use handle_* naming pattern def handle_info(self, **kwargs) -> dict: return {"version": "1.0", "status": "ready"} ``` ## Configuration Reference **model.yaml:** ```yaml name: my_model # Required: unique model name version: "1.0.0" # Model version python: ">=3.11" # Python version requirement client: module: model # Python module name (model.py) class: MyModelClient # Class name in module requirements: requirements.txt # Dependencies file handlers: # Optional: document handlers - name: predict input: x: float output: result: float health: interval: 30 # Health check interval (seconds) timeout: 5 # Health check timeout max_failures: 3 # Failures before restart resources: # Optional: resource limits memory_limit: 4G cpu_limit: 2 gpu_ids: [0, 1] # CUDA_VISIBLE_DEVICES ``` ## License MIT License - see [LICENSE](LICENSE) for details.
text/markdown
null
Johanna <johanna@example.com>
null
null
null
ipc, machine-learning, model-serving, pytorch, subprocess, tensorflow
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: POSIX", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Topic :: Scientific/Engineering :: Ar...
[]
null
null
>=3.11
[]
[]
[]
[ "click>=8.0", "pydantic>=2.0", "pyyaml>=6.0", "rich>=13.0", "django>=4.2; extra == \"all\"", "mypy>=1.0; extra == \"all\"", "pytest-asyncio>=0.21; extra == \"all\"", "pytest-cov>=4.0; extra == \"all\"", "pytest>=7.0; extra == \"all\"", "ruff>=0.1.0; extra == \"all\"", "mypy>=1.0; extra == \"dev\...
[]
[]
[]
[ "Homepage, https://github.com/johanna/pymodelserve", "Documentation, https://github.com/johanna/pymodelserve#readme", "Repository, https://github.com/johanna/pymodelserve" ]
uv/0.9.2
2026-02-19T12:12:36.045464
pymodelserve-0.1.1.tar.gz
95,557
a1/25/098a845ba0099c1a394c8bee96f6a89bf76ebd4d5d166f73c6f59d3d29e9/pymodelserve-0.1.1.tar.gz
source
sdist
null
false
14da7569247750fddfea357010399c2d
470f257ef5ad432ddcd6ddb912144d9bc4d025dc770faa1ad0b7fd7501de3490
a125098a845ba0099c1a394c8bee96f6a89bf76ebd4d5d166f73c6f59d3d29e9
MIT
[ "LICENSE" ]
243
2.1
usbadc10
1.0.2
usbadc10 protocol python binding.
## USBADC10 This is a python binding for usbadc10 cross-platform library for USBADC10 - a device that converts an input analog signal into a discrete code, includes 10 channels of a 12—bit ADC, an STM32 microcontroller and a USB interface that supplies power and reads digitized data. ![image](https://raw.githubusercontent.com/EPC-MSU/UALab/3982c42a179a38d2dce1af6235d4ec88ee8cab51/media/usbadc10_board.jpg) ### Installation ``` pip install usbadc10 ``` ### Minimal example ```python from usbadc10 import Usbadc10DeviceHandle # Set correct device URI here # Format for Windows: com:\\.\COM5 # Format for Linux: /dev/ttyACM0 # Format for MacOS: com:///dev/tty.usbmodem000001234 device_uri = r'com:\\.\COM5' # Create and open device instance device = Usbadc10DeviceHandle(device_uri) # Read raw data raw_data_all_channels = list(device.get_conversion_raw().data) print("List of raw ADC counts from all channels:\n", raw_data_all_channels) # Read voltages in (10 * mV) units voltage_all_channels = list(device.get_conversion().data) print("List of voltages from all channels (in 10 * mV):\n", voltage_all_channels) # Convert measurements to mV voltage_all_channels_mV = [value/10 for value in voltage_all_channels] print("List of voltages from all channels (in mV):\n", voltage_all_channels_mV) # Close the device device.close_device() ``` ### More information * usbadc10 website: https://usbadc10.physlab.ru/
text/markdown
null
EPC MSU <info@physlab.ru>
null
null
Copyright (c) 2024 EPC MSU Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
null
[ "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: Microsoft :: Windows :: Windows 10", "Operating System :: Microsoft :: Windows :: Windows 11", "Operating System :: Microsoft :: Windows :: Windows 7", "Operating System :: Microsoft :: Windows :: Windows 8", "Op...
[]
null
null
>=3.6
[]
[]
[]
[]
[]
[]
[]
[]
twine/6.2.0 CPython/3.9.6
2026-02-19T12:12:29.592102
usbadc10-1.0.2.tar.gz
3,198,565
c0/d8/b63f7bc13e821523b0ec31ca29a16023f891314fb851084cd8ceeb8fe591/usbadc10-1.0.2.tar.gz
source
sdist
null
false
de38c07738987999479ea551de173fb1
c74a533b5484314f273e06e50b18582b707ea2bf16543f75d546eeff8cfeae00
c0d8b63f7bc13e821523b0ec31ca29a16023f891314fb851084cd8ceeb8fe591
null
[]
245
2.4
zExceptions
6.0
zExceptions contains common exceptions used in Zope.
Overview ======== zExceptions contains common exceptions and helper functions related to exceptions as used in Zope. Changelog ========= 6.0 (2026-02-19) ---------------- - Add support for Python 3.12, 3.13, and 3.14. - Drop support for Python 3.7, 3.8, and 3.9. 5.0 (2023-06-28) ---------------- - Drop support for Python 2.7, 3.5, 3.6. 4.3 (2022-12-18) ---------------- - Sort imports with isort. - Add support for Python 3.11. 4.2 (2021-10-22) ---------------- - Add support for Python 3.8, 3.9 and 3.10. - Drop support for Python 3.4. - Add an exception for the HTTP status code 418. - Don't override the `content-type` header if already set. (`#12 <https://github.com/zopefoundation/zExceptions/pull/12>`_) 4.1 (2018-10-05) ---------------- - Add support for Python 3.7. 4.0 (2018-01-27) ---------------- - Drop support for string exceptions. 3.6.1 (2017-05-17) ------------------ - Increase Python 3 compatibility 3.6 (2017-02-05) ---------------- - Add realm as an argument to unauthorized exceptions, its presence causing a `WWW-Authenticate` header to be emitted. - Set `location` header during `__init__` of redirect exceptions. 3.5 (2017-02-05) ---------------- - Drop support for Python 3.3, add support for Python 3.6. - Use `str(self)` as detail if it is not set. - Add a `setHeader` method to add a response header to an HTTPException. - `upgradeException` now also supports finding an HTTPException class with the same name as a non-HTTPException class. 3.4 (2016-09-08) ---------------- - Use `HTTPException.body_template` when title and detail are set. - Add new title and detail attributes to HTTPException. 3.3 (2016-08-06) ---------------- - Add exception classes for all HTTP status codes. 3.2 (2016-07-22) ---------------- - Implement basic subset of Response features in HTTPException class. 3.1 (2016-07-22) ---------------- - Mark exceptions with appropriate zope.publisher interfaces. - Add a new common base class `zExceptions.HTTPException` to all exceptions. 3.0 (2016-04-03) ---------------- - Add compatibility with PyPy and Python 3. - Arguments to the Unauthorized exception are assumed to be utf8-encoded if they are bytes. 2.13.0 (2010-06-05) ------------------- - Released as separate package.
text/x-rst
Zope Foundation and Contributors
zope-dev@zope.dev
null
null
ZPL-2.1
null
[ "Development Status :: 6 - Mature", "Environment :: Web Environment", "Framework :: Zope :: 5", "License :: OSI Approved :: Zope Public License", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "...
[]
https://github.com/zopefoundation/zExceptions
null
>=3.10
[]
[]
[]
[ "setuptools", "zope.interface", "zope.publisher", "zope.security" ]
[]
[]
[]
[ "Issue Tracker, https://github.com/zopefoundation/zExceptions/issues", "Sources, https://github.com/zopefoundation/zExceptions" ]
twine/6.1.0 CPython/3.12.12
2026-02-19T12:12:15.426127
zexceptions-6.0.tar.gz
16,204
56/2b/66ad455cbf26e6486bfb00bfa9d854531e02b96402e59500d852693b1884/zexceptions-6.0.tar.gz
source
sdist
null
false
da252edc9af59e4370d6baabbb6383e7
2785efab97362e57fa9e394a15c31a2e481b08fa953417d34f09eb1edc201242
562b66ad455cbf26e6486bfb00bfa9d854531e02b96402e59500d852693b1884
null
[ "LICENSE.txt" ]
0
2.4
analysim-jupyterlite-integration
0.1.0
A JupyterLite frontend extension that enables seamless integration between JupyterLite and Analysim.
# analysim_jupyterlite_integration [![Github Actions Status](/workflows/Build/badge.svg)](/actions/workflows/build.yml) A JupyterLite frontend extension that enables seamless integration between JupyterLite and Analysim. ## Requirements - JupyterLab >= 4.0.0 ## Install To install the extension, execute: ```bash pip install analysim_jupyterlite_integration ``` ## Uninstall To remove the extension, execute: ```bash pip uninstall analysim_jupyterlite_integration ``` ## Contributing ### Development install Note: You will need NodeJS to build the extension package. The `jlpm` command is JupyterLab's pinned version of [yarn](https://yarnpkg.com/) that is installed with JupyterLab. You may use `yarn` or `npm` in lieu of `jlpm` below. ```bash # Clone the repo to your local environment # Change directory to the analysim_jupyterlite_integration directory # Set up a virtual environment and install package in development mode python -m venv .venv source .venv/bin/activate pip install --editable "." # Link your development version of the extension with JupyterLab jupyter labextension develop . --overwrite # Rebuild extension Typescript source after making changes # IMPORTANT: Unlike the steps above which are performed only once, do this step # every time you make a change. jlpm build ``` You can watch the source directory and run JupyterLab at the same time in different terminals to watch for changes in the extension's source and automatically rebuild the extension. ```bash # Watch the source directory in one terminal, automatically rebuilding when needed jlpm watch # Run JupyterLab in another terminal jupyter lab ``` With the watch command running, every saved change will immediately be built locally and available in your running JupyterLab. Refresh JupyterLab to load the change in your browser (you may need to wait several seconds for the extension to be rebuilt). By default, the `jlpm build` command generates the source maps for this extension to make it easier to debug using the browser dev tools. To also generate source maps for the JupyterLab core extensions, you can run the following command: ```bash jupyter lab build --minimize=False ``` ### Development uninstall ```bash pip uninstall analysim_jupyterlite_integration ``` In development mode, you will also need to remove the symlink created by `jupyter labextension develop` command. To find its location, you can run `jupyter labextension list` to figure out where the `labextensions` folder is located. Then you can remove the symlink named `analysim-jupyterlite-integration` within that folder. ### Packaging the extension See [RELEASE](RELEASE.md)
text/markdown
null
mohab <mohab_sobhy@outlook.com>
null
null
BSD 3-Clause License Copyright (c) 2026, mohab All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: 1. Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. 2. Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. 3. Neither the name of the copyright holder nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
jupyter, jupyterlab, jupyterlab-extension
[ "Framework :: Jupyter", "Framework :: Jupyter :: JupyterLab", "Framework :: Jupyter :: JupyterLab :: 4", "Framework :: Jupyter :: JupyterLab :: Extensions", "Framework :: Jupyter :: JupyterLab :: Extensions :: Prebuilt", "License :: OSI Approved :: BSD License", "Programming Language :: Python", "Prog...
[]
null
null
>=3.10
[]
[]
[]
[]
[]
[]
[]
[ "Homepage, https://github.com/YOUR_USERNAME/analysim_jupyterlite_integration", "Repository, https://github.com/YOUR_USERNAME/analysim_jupyterlite_integration", "Issues, https://github.com/YOUR_USERNAME/analysim_jupyterlite_integration/issues" ]
twine/6.2.0 CPython/3.14.2
2026-02-19T12:11:45.020711
analysim_jupyterlite_integration-0.1.0.tar.gz
87,037
d6/df/161e63faa19d6e94a64b49f9b4d1cbef5d6fd1f0128098cc72844b6a2c76/analysim_jupyterlite_integration-0.1.0.tar.gz
source
sdist
null
false
3770415e2139733864046efc2191cc1f
40df94df3a469300bf133ee485152ffa610e5506b720829cd6e525e2f8c94f1e
d6df161e63faa19d6e94a64b49f9b4d1cbef5d6fd1f0128098cc72844b6a2c76
null
[ "LICENSE" ]
261
2.1
odoo-addon-purchase-product-internal-reference
18.0.1.0.0.1
Display internal product reference in purchase orders
=================================== Purchase product internal reference =================================== .. !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! !! This file is generated by oca-gen-addon-readme !! !! changes will be overwritten. !! !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! !! source digest: sha256:ceff71185f640e677aa77a0d5c04d80ec0b0be79584127de3acff5dd7894b8b7 !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! .. |badge1| image:: https://img.shields.io/badge/maturity-Beta-yellow.png :target: https://odoo-community.org/page/development-status :alt: Beta .. |badge2| image:: https://img.shields.io/badge/licence-AGPL--3-blue.png :target: http://www.gnu.org/licenses/agpl-3.0-standalone.html :alt: License: AGPL-3 .. |badge3| image:: https://img.shields.io/badge/github-sygel--technology%2Fsy--purchase--workflow-lightgray.png?logo=github :target: https://github.com/sygel-technology/sy-purchase-workflow/tree/18.0/purchase_product_internal_reference :alt: sygel-technology/sy-purchase-workflow |badge1| |badge2| |badge3| This module allows configuring the Purchase application to display the internal product reference (``default_code``) in the product selection dropdown. By default, when adding a product to a document for a partner marked as a supplier, Odoo may display the supplier reference defined in ``product.supplierinfo`` instead of the internal reference. When the option *Show internal reference* is enabled at company level, the product selector dropdown always shows the internal reference instead of the supplier reference. **Table of contents** .. contents:: :local: Installation ============ To install this module, you need to: - Only install Configuration ============= This setting must be configured per company: 1. Go to **Purchase** → **Settings** 2. Enable **"Show internal product reference"** 3. Save changes Usage ===== 1. Go to **Purchase** → **Orders** → **Requests for Quotation** 2. Create a new RFQ 3. Select a vendor 4. Add a product line When the option is enabled: - The product selector dropdown shows the internal reference. Bug Tracker =========== Bugs are tracked on `GitHub Issues <https://github.com/sygel-technology/sy-purchase-workflow/issues>`_. In case of trouble, please check there if your issue has already been reported. If you spotted it first, help us to smash it by providing a detailed and welcomed `feedback <https://github.com/sygel-technology/sy-purchase-workflow/issues/new?body=module:%20purchase_product_internal_reference%0Aversion:%2018.0%0A%0A**Steps%20to%20reproduce**%0A-%20...%0A%0A**Current%20behavior**%0A%0A**Expected%20behavior**>`_. Do not contact contributors directly about support or help with technical issues. Credits ======= Authors ------- * Sygel Contributors ------------ - `Sygel <https://www.sygel.es>`__: - Ángel García de la Chica Herrera - Valentín Vinagre - Ángel Rivas Maintainers ----------- This module is part of the `sygel-technology/sy-purchase-workflow <https://github.com/sygel-technology/sy-purchase-workflow/tree/18.0/purchase_product_internal_reference>`_ project on GitHub. You are welcome to contribute.
text/x-rst
Sygel, Odoo Community Association (OCA)
support@odoo-community.org
null
null
AGPL-3
null
[ "Programming Language :: Python", "Framework :: Odoo", "Framework :: Odoo :: 18.0", "License :: OSI Approved :: GNU Affero General Public License v3" ]
[]
https://github.com/sygel-technology/sy-purchase-workflow
null
>=3.10
[]
[]
[]
[ "odoo==18.0.*" ]
[]
[]
[]
[]
twine/5.1.1 CPython/3.12.3
2026-02-19T12:11:32.891874
odoo_addon_purchase_product_internal_reference-18.0.1.0.0.1-py3-none-any.whl
39,450
a9/2a/51537a35ec18e713f5876e52d1fcff345715e875477a172c16cf8384fd54/odoo_addon_purchase_product_internal_reference-18.0.1.0.0.1-py3-none-any.whl
py3
bdist_wheel
null
false
0bfb10591d6a7ff2b03e169f1db01c84
0c3ae7591c23c0e9b0c20158a4dbebc9ffaea82f425fa6d37c6e1d1e99409abc
a92a51537a35ec18e713f5876e52d1fcff345715e875477a172c16cf8384fd54
null
[]
101
2.4
cogstim
0.7.1
Visual cognitive-stimulus generator (ANS dots, shapes, gratings)
# CogStim – Visual Cognitive-Stimulus Generator [![PyPI version](https://img.shields.io/pypi/v/cogstim.svg)](https://pypi.org/project/cogstim/) [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/licenses/MIT) [![CI](https://github.com/eudald-seeslab/cogstim/workflows/CI/badge.svg)](https://github.com/eudald-seeslab/cogstim/actions) [![Coverage Status](https://coveralls.io/repos/github/eudald-seeslab/cogstim/badge.svg?branch=main)](https://coveralls.io/github/eudald-seeslab/cogstim?branch=main) CogStim is a small Python toolkit that produces **synthetic image datasets** commonly used in cognitive–neuroscience and psychophysics experiments, such as: * Shape discrimination (e.g. *circle vs star*). * Colour discrimination (e.g. *yellow vs blue* circles). * Approximate Number System (ANS) dot arrays with two colours. * Single-colour dot arrays for number-discrimination tasks. * Custom combinations of geometrical *shapes × colours*. * Rotated stripe patterns ("lines" dataset) for orientation discrimination. * Fixation targets (A, B, C, AB, AC, BC, ABC) with configurable colours. All stimuli are generated as PNG files with a default size of 512 × 512 pixels (configurable via `--img-size`). ## Installation ```bash pip install cogstim ``` ## Documentation - **[Quick Start](docs/index.md)** – Installation and first steps - **[User Guide](docs/guide.md)** – Detailed documentation for each task - **[Recipes](docs/recipes.md)** – Copy-paste commands for common goals - **[FAQ](docs/faq.md)** – Troubleshooting and common questions ### For LLM/AI Agents - **[LLM Documentation](docs/LLM_DOCUMENTATION.md)** – Single-file comprehensive documentation optimized for feeding to Large Language Models (Context + Architecture + API Reference) ## Command-line interface CogStim provides a simple command-line interface with task-specific subcommands: ```bash cogstim <task> [options] ``` Available tasks: - `shapes` – Shape discrimination (e.g., circles vs stars) - `colours` – Colour discrimination (same shape, different colours) - `ans` – Two-colour dot arrays (Approximate Number System) - `one-colour` – Single-colour dot arrays (quantity discrimination) - `match-to-sample` – Match-to-sample dot array pairs - `lines` – Rotated stripe/line patterns - `fixation` – Fixation target images - `custom` – Custom shape/colour combinations For help on a specific task: ```bash cogstim <task> --help ``` ### Common options Most tasks accept these options: - `--train-num N` – Number of training image sets (default: 10) - `--test-num N` – Number of test image sets (default: 0) - `--output-dir PATH` – Output directory (default: `images/<task>`) - `--img-size SIZE` – Image size in pixels (default: 512) - `--background-colour COLOUR` – Background colour (default: white) - `--seed SEED` – Random seed for reproducible generation - `--demo` – Generate a quick preview with 8 training images > **Note**: `--train-num` and `--test-num` refer to the number of image _sets_ created. An image set is a group of images that combines all the possible parameter combinations. For shapes and colours, an image set is about 200 images, whereas for ANS it's around 75 images, depending on the parameters. > **Note**: All CLI arguments use British spelling. > **Note**: Use `--seed SEED` (where SEED is an integer) to make generation deterministic and reproducible. Without a seed, each run will produce different random variations. ### Task specification from CSV For some tasks, you can specify exactly which stimuli to generate via a CSV file instead of using ratios or parameter ranges. You can do this with the `--tasks-csv PATH` to point to a CSV (find below the format of the CSV depending on the task). This method allows to create "n" copies of the tasks in the CSV (`--tasks-copies N` (default: 1) to repeat the distribution N times). When `--tasks-csv` is set, `--ratios` and `--min-point-num` / `--max-point-num` are ignored. For now, this is supported only for the following tasks, with the CSV specifications | Task | CSV columns | |------------------|-----------------------------| | Match to sample | sample, match, equalized | ## Examples ### Shape recognition – *circle vs star* in yellow ```bash cogstim shapes --train-num 60 --test-num 20 ``` For reproducible results, add the `--seed` option: ```bash cogstim shapes --train-num 60 --test-num 20 --seed 1234 ``` <table><tr> <td><img src="https://raw.githubusercontent.com/eudald-seeslab/cogstim/main/assets/examples/circle.png" alt="Yellow circle" width="220"/></td> <td><img src="https://raw.githubusercontent.com/eudald-seeslab/cogstim/main/assets/examples/star.png" alt="Yellow star" width="220"/></td> </tr></table> ### Colour recognition – yellow vs blue circles (no positional jitter) ```bash cogstim colours --train-num 60 --test-num 20 --no-jitter ``` <table><tr> <td><img src="https://raw.githubusercontent.com/eudald-seeslab/cogstim/main/assets/examples/circle.png" alt="Yellow circle" width="220"/></td> <td><img src="https://raw.githubusercontent.com/eudald-seeslab/cogstim/main/assets/examples/circle_blue.png" alt="Blue circle" width="220"/></td> </tr></table> ### Approximate Number System (ANS) dataset with easy ratios only ```bash cogstim ans --ratios easy --train-num 100 --test-num 40 ``` <table><tr> <td><img src="https://raw.githubusercontent.com/eudald-seeslab/cogstim/main/assets/examples/ans_equalized.png" alt="ANS equalized" width="220"/></td> <td><img src="https://raw.githubusercontent.com/eudald-seeslab/cogstim/main/assets/examples/ans.png" alt="ANS non-equalized" width="220"/></td> </tr></table> > Note that on the left image, total surfaces are equalized, and, on the right image, dot size is random. This is based on Halberda et al. (2008). ### Match-to-sample (MTS) – dot arrays (sample/match) with controlled total surface ```bash cogstim match-to-sample \ --ratios easy \ --train-num 50 --test-num 20 \ --min-point-num 1 --max-point-num 10 \ --dot-colour yellow ``` - Generates pairs of images per trial: `*_s.png` (sample) and `*_m.png` (match). - For half of the trials, total dot surface is equalized between sample and match; for the other half, dot sizes are random. - The target total surface for the match is derived from the sample image of the same trial. - Unequal pairs are built from the same ratio set used by ANS, with both orders (n→m and m→n) included, and equal (n=m) trials added to balance labels. - Output layout: `images/match_to_sample/{train|test}/img_{n}_{m}_{k}[...]_s.png` and corresponding `img_{n}_{m}_{k}[...]_m.png`. This task is based on Sella et al. (2013). ### Single-colour dot arrays numbered 1-5, total surface area held constant ```bash cogstim one-colour --train-num 50 --test-num 20 --min-point-num 1 --max-point-num 5 ``` <table><tr> <td><img src="https://raw.githubusercontent.com/eudald-seeslab/cogstim/main/assets/examples/dots_two.png" alt="Two circles" width="220"/></td> <td><img src="https://raw.githubusercontent.com/eudald-seeslab/cogstim/main/assets/examples/dots_five.png" alt="Five circles" width="220"/></td> </tr></table> ### Custom dataset – green/red triangles & squares ```bash cogstim custom --shapes triangle square --colours red green --train-num 50 --test-num 20 ``` <table><tr> <td><img src="https://raw.githubusercontent.com/eudald-seeslab/cogstim/main/assets/examples/triangle_red.png" alt="Red triangle" width="220"/></td> <td><img src="https://raw.githubusercontent.com/eudald-seeslab/cogstim/main/assets/examples/square_green.png" alt="Green square" width="220"/></td> </tr></table> ### Lines dataset – rotated stripe patterns ```bash cogstim lines --train-num 50 --test-num 20 --angles 0 45 90 135 --min-stripes 3 --max-stripes 5 ``` <table><tr> <td><img src="https://raw.githubusercontent.com/eudald-seeslab/cogstim/main/assets/examples/lines_vertical.png" alt="Vertical lines" width="220"/></td> <td><img src="https://raw.githubusercontent.com/eudald-seeslab/cogstim/main/assets/examples/lines_horizontal.png" alt="Horizontal lines" width="220"/></td> </tr></table> This task is based on Srinivasan (2021). ### Fixation targets – A/B/C/AB/AC/BC/ABC ```bash cogstim fixation \ --all-types \ --background-colour black --symbol-colour white \ --img-size 512 --dot-radius-px 6 --disk-radius-px 128 --cross-thickness-px 24 \ --cross-arm-px 128 ``` - The symbol uses a single colour (`--symbol-colour`). - Composite types BC/ABC are rendered by overdrawing the cross and/or central dot with the background colour to create cut-outs, matching the figure convention in Thaler et al. (2013). - For fixation targets, exactly one image is generated per type. - Use `--all-types` to generate all seven types; otherwise, choose a subset via `--types`. - Control cross bar length using `--cross-arm-px` (half-length from center), and thickness via `--cross-thickness-px`. Output folder layout for fixation targets: ``` images/fixation/ ``` These shapes are based on Thaler et al. (2013). They recommend using ABC. <img src="https://raw.githubusercontent.com/eudald-seeslab/cogstim/main/assets/examples/fix_ABC.png" alt="Fixation point example" width="220"/> ## Output The generated folder structure is organised by *phase / class*, e.g. ``` images/two_shapes/ ├── train/ │ ├── circle/ │ └── star/ └── test/ ├── circle/ └── star/ ``` ## License This project is distributed under the **MIT License** – see the `LICENCE` file for details. ## TODO's - The equalization algorithm of match-to-sample could be improved. - Extend CSV-based task specification to other tasks (ANS, one-colour, shapes, etc.). - Check that the image is big enough for the parameters set. ## References - Halberda, J., Mazzocco, M. M. M., & Feigenson, L. (2008). Individual differences in non-verbal number acuity correlate with maths achievement. Nature, 455(7213), 665-668. https://doi.org/10.1038/nature07246 - Sella, F., Lanfranchi, S., & Zorzi, M. (2013). Enumeration skills in Down syndrome. Research in Developmental Disabilities, 34(11), 3798-3806. https://doi.org/10.1016/j.ridd.2013.07.038 - Srinivasan, M. V. (2021). Vision, perception, navigation and ‘cognition’ in honeybees and applications to aerial robotics. Biochemical and Biophysical Research Communications, 564, 4-17. https://doi.org/10.1016/j.bbrc.2020.09.052 - Thaler, L., Schütz, A. C., Goodale, M. A., & Gegenfurtner, K. R. (2013). What is the best fixation target? The effect of target shape on stability of fixational eye movements. Vision Research, 76, 31–42. https://doi.org/10.1016/j.visres.2012.10.012
text/markdown
null
Eudald Correig-Fraga <eudald.correig@urv.cat>
null
null
The MIT License (MIT) Copyright (c) <year> Adam Veldhousen Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
null
[ "Development Status :: 3 - Alpha", "Intended Audience :: Science/Research", "Programming Language :: Python :: 3 :: Only", "Topic :: Scientific/Engineering :: Artificial Intelligence" ]
[]
null
null
>=3.10
[]
[]
[]
[ "Pillow>=9.0.0", "numpy>=1.21.0", "tqdm>=4.60.0", "pytest; extra == \"dev\"", "pytest-cov; extra == \"dev\"", "coverage; extra == \"dev\"", "coveralls; extra == \"dev\"", "black; extra == \"dev\"", "ruff; extra == \"dev\"", "mkdocs-material; extra == \"docs\"" ]
[]
[]
[]
[ "Homepage, https://github.com/eudald-seeslab/cogstim", "BugTracker, https://github.com/eudald-seeslab/cogstim/issues" ]
twine/6.2.0 CPython/3.12.10
2026-02-19T12:10:44.484188
cogstim-0.7.1.tar.gz
62,072
2a/16/a8ee89fba9320caf1b63333bde4d9a1b934ba1e06c4030630a0b60504338/cogstim-0.7.1.tar.gz
source
sdist
null
false
60488d59c4132e8ae409712bee732046
4cc97163a475b399345a636bb2bcb9cf1b54a00eebbe450c4ddaef87512b6f7f
2a16a8ee89fba9320caf1b63333bde4d9a1b934ba1e06c4030630a0b60504338
null
[ "LICENCE" ]
255
2.4
codestory-cli
0.2.2
Every project should have a good story behind it. With a good story (git history), you can reason about changes in a way that would be impossible just looking at a codebase. But let us do the work in creating that story
# <img src="resources/icon.svg" width="32" height="32" align="center" /> Codestory CLI ![Tests](https://img.shields.io/github/actions/workflow/status/CodeStoryBuild/CodeStoryCli/pytest.yml?label=tests) ![Docs Deployment](https://img.shields.io/github/actions/workflow/status/CodeStoryBuild/CodeStoryCli/deploy-docs.yml?label=docs) ![GitHub Release](https://img.shields.io/github/v/release/CodeStoryBuild/CodeStoryCli?include_prereleases&label=release) ![PyPI Version](https://img.shields.io/pypi/v/codestory-cli?color=blue&label=pypi&cacheSeconds=86400) ![PyPI Downloads](https://img.shields.io/pypi/dm/codestory-cli?color=blue&label=downloads&cacheSeconds=86400) The Codestory CLI is a high level interface over git. Why? Git is a powerful tool. It is also an excellent tool for shooting yourself in the foot. While Git provides the infrastructure for version control, you rarely need to be down in the pipes for daily work. Codestory lets you do 99% of regular tasks much more effectively. When you need Git, its always there for those more complex tasks. It does not replace Git; it allows you to use it effectively without the headache. Codestory provides powerful commands to enhance your version control workflow: commit changes into small, logical commits with AI-generated messages; fix past commits by intelligently splitting them into more focused changes; clean entire repository histories to improve commit quality; and configure settings to tailor the tool to your needs. Think of this as a natural transition to a higher-level version control, where you can worry less about housekeeping and focus on building. ## Getting Started For detailed instructions on how to install and use the Codestory CLI, visit our **[getting started](https://cli.codestory.build/docs/getting-started/)** page. ## Documentation For comprehensive guides, configuration priority, and supported languages, visit our **[docs](https://cli.codestory.build/docs)** page. The documentation site provides detailed information on: - [Usage Guides](https://cli.codestory.build/docs/usage/) - [Configuration Scopes (Global vs. Local vs. Env)](https://cli.codestory.build/docs/configuration/) - [Supported Model Providers and Languages](https://cli.codestory.build/docs/reference/supported/) - [Core Design and Architecture](https://cli.codestory.build/docs/design/) ## Main-Cleaned For a real example of what codestory can do, checkout the `main-cleaned` branch of this repository. It is a cleaned up version of the main branch of this repository, created using the `cst clean` command. ## Contributing If you find a bug, feel free to create a pull request. ## License The code is licensed under GPLv2, just like Git.
text/markdown
null
Adem Can <ademfcan@gmail.com>
null
null
null
null
[]
[]
null
null
>=3.11
[]
[]
[]
[ "typer-slim==0.20.0", "colorama==0.4.6", "loguru==0.7.3", "tree-sitter==0.25.2", "tree-sitter-language-pack==0.13.0", "platformdirs==4.5.0", "tqdm==4.67.1", "fastembed>=0.7.4", "scikit-learn<1.9.0,>=1.8.0", "networkx<4.0,>=3.0", "python-dotenv>=1.2.1", "aisuite-extendedmodels[anthropic,aws,azu...
[]
[]
[]
[]
twine/6.2.0 CPython/3.13.12
2026-02-19T12:10:03.690172
codestory_cli-0.2.2.tar.gz
61,399,826
6c/56/ecf460abd1ceb70f5db2d889e84a535c054a856ca9a060a04ef397b06e26/codestory_cli-0.2.2.tar.gz
source
sdist
null
false
ac7c787c57cbb6a7bd2ff78c4cad0b18
cbe3e52a2894c2e775e31706c5c9deae321084d25ac5797ca8e0bcbd5892859d
6c56ecf460abd1ceb70f5db2d889e84a535c054a856ca9a060a04ef397b06e26
GPL-2.0-only
[]
270
2.1
ellipsis
3.3.6
Package to interact with the Ellipsis API
# Ellipsis Drive Python Package This package is meant to help you interact with the Ellipsis API. You can install this package using `pip install ellipsis` For documentation see https://ellipsis-package.readthedocs.io This package is meant to ease the use of the Ellipsis Drive API in your Python projects. # Examples Below are some code examples. import ellipsis as el # log in token = el.account.logIn("username", "password") # retrieve all maps in "My Drive" maps = el.account.listRoot("myDrive", pathType='layer', token=token) Another example import ellipsis as el folderId = '46e1e919-8b73-42a3-a575-25c6d45fd93b' token = el.account.logIn("username", "password") info = el.path.get(folderId, token) layers = el.path.listPath(folderId, pathType='layer', token = token, listAll = True) folders = el.path.listPath(folderId, pathType='folder', token = token, listAll = True)
text/markdown
Daniel van der Maas
daniel@ellipsis-drive.com
null
null
null
null
[ "Programming Language :: Python :: 3", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent" ]
[]
https://github.com/ellipsis-drive-internal/python-package
null
>=3.6
[]
[]
[]
[]
[]
[]
[]
[]
twine/5.0.0 CPython/3.12.3
2026-02-19T12:10:00.766460
ellipsis-3.3.6.tar.gz
46,729
4a/a0/5650dd44961e8d8584c27d9d33538d6c54ebc59855f959d70dc9414eed6e/ellipsis-3.3.6.tar.gz
source
sdist
null
false
b3c7d4f9b819f7656556231888cdc463
40563605f625c6c279481175b240c6a706072bd42543326d4c5166715508d058
4aa05650dd44961e8d8584c27d9d33538d6c54ebc59855f959d70dc9414eed6e
null
[]
275
2.4
symphony_bdk_python
2.11.2
Symphony Bot Development Kit for Python
[![FINOS - Active](https://cdn.jsdelivr.net/gh/finos/contrib-toolbox@master/images/badge-active.svg)](https://community.finos.org/docs/governance/Software-Projects/stages/active) [![License](https://img.shields.io/badge/License-Apache%202.0-blue.svg)](https://opensource.org/licenses/Apache-2.0) [![Python](https://img.shields.io/badge/python-3.9%20%7C%203.10%20%7C%203.11%20%7C%203.12%20%7C%203.13-blue)](https://www.python.org/downloads/) [![Pypi](https://img.shields.io/pypi/v/symphony-bdk-python)](https://pypi.org/project/symphony-bdk-python/) ![GitHub Workflow Status (branch)](https://img.shields.io/github/workflow/status/finos/symphony-bdk-python/build/main) # Symphony BDK for Python This is the Symphony BDK for Python to help develop bots and interact with the [Symphony REST APIs](https://rest-api.symphony.com). ## Project Overview Symphony BDK for Python provides tools for building bots and integrating with Symphony APIs. This document outlines its usage, installation, and contribution guidelines. ## Prerequisites - Python 3.9 or higher - [Poetry](https://python-poetry.org/docs/#installation) ## Installation Steps ### Option 1: Build from Source - Install dependencies: `poetry install` - Build the package: `poetry build` - Run tests: `poetry run pytest` - Perform a lint scan locally: `poetry run ruff check .` - Format code locally: `poetry run ruff format .` - Generate documentation locally: `cd docsrc && make html` ### Setting Up Git Hooks This project uses `pre-commit` with `ruff` to automatically format and lint code. This is the recommended setup for contributors to ensure code style consistency. 1. **Install development dependencies** (this will also install `pre--commit` and `ruff`): ```bash poetry install ``` 2. **Install the git hooks**: ```bash poetry run pre-commit install ``` Now, `ruff` will automatically run on every commit, formatting your code and checking for linting errors. ### Verification Verify the successful installation by running any of the following commands: ``` poetry --version ``` ## External Documents Refer to the following for additional guidance: - [Reference Documentation](https://symphony-bdk-python.finos.org/) - [Getting Started Guide](https://symphony-bdk-python.finos.org/markdown/getting_started.html) ## Roadmap The next milestone is the [2.5.x](https://github.com/finos/symphony-bdk-python/milestone/6), focused on delivering improvements and bug fixes. ## Contributing To contribute: 1. Fork the repository. 2. Create a feature branch: `git checkout -b feature/fooBar` 3. Read the [Contribution Guidelines](CONTRIBUTING.md) and [Community Code of Conduct](https://www.finos.org/code-of-conduct) 4. Commit changes: `git commit -am 'Add some fooBar'` 5. Push changes: `git push origin feature/fooBar` 6. Open a Pull Request. _NOTE:_ Ensure you have an active Individual Contributor License Agreement (ICLA) or Corporate Contribution License Agreement (CCLA) with FINOS. For further inquiries, email [help@finos.org](mailto:help@finos.org). ### Updating Generated Code Python BDK uses [OpenAPITools/openapi-generator](https://github.com/OpenAPITools/openapi-generator/) to generate code. To update the generated code, follow these steps: 1. Checkout the latest branch of the fork (e.g., [sym-python-5.5.0](https://github.com/SymphonyPlatformSolutions/openapi-generator/tree/sym-python-5.5.0)). 2. Update the fork source code, review, and merge it. 3. Generate the JAR file in `openapi-generatormodules/openapi-generator-cli/target/openapi-generator-cli.jar`: - Use Maven: ```bash mvn clean install -Dmaven.test.skip=true && mvn clean package -Dmaven.test.skip=true ``` - Alternatively, use IntelliJ's build button to build the project and generate the JAR file. 4. Copy the JAR file to the Python BDK repository: `symphony-api-client-python/api_client_generation/openapi-generator-cli.jar`. 5. Execute the generation script: ```bash ./generate.sh ``` 6. Commit and push the newly generated code along with the updated JAR file. ## License Copyright 2021 Symphony LLC Distributed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0). SPDX-License-Identifier: [Apache-2.0](https://spdx.org/licenses/Apache-2.0).
text/markdown
Symphony Platform Solutions
symphony@finos.org
null
null
Apache-2.0
null
[ "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14" ]
[]
https://github.com/finos/symphony-bdk-python
null
!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,!=3.8.*,>=3.9
[]
[]
[]
[ "nulltype<3.0.0,>=2.3.1", "python-dateutil<3.0.0,>=2.8.2", "urllib3<3,>=2.5.0", "aiohttp<4.0.0,>=3.12.14", "pyyaml<7.0,>=6.0", "PyJWT<3.0.0,>=2.10.0", "cryptography<47.0.0,>=46.0.0", "tenacity<9.0.0,>=8.0.1", "defusedxml<0.8.0,>=0.7.1", "docutils==0.16" ]
[]
[]
[]
[ "Repository, https://github.com/finos/symphony-bdk-python", "Documentation, https://symphony-bdk-python.finos.org/" ]
poetry/2.2.1 CPython/3.9.25 Linux/6.14.0-1017-azure
2026-02-19T12:09:51.447778
symphony_bdk_python-2.11.2.tar.gz
326,423
d5/72/7ead0b321bc310b5e61b1c993bfce8c963714beb1fabfdaa898e8633b93a/symphony_bdk_python-2.11.2.tar.gz
source
sdist
null
false
44bd24c60ab63faea651eaef166ccb27
667d6635432a6990a419ea24579a50e30069565e5534dc8c79cd7edc19306c52
d5727ead0b321bc310b5e61b1c993bfce8c963714beb1fabfdaa898e8633b93a
null
[]
0
2.4
vid2cc-ai
0.1.5
AI-powered subtitle generator using Whisper and FFmpeg
# vid2cc-AI 🎙️🎬 [![PyPI version](https://img.shields.io/pypi/v/vid2cc-ai.svg)](https://pypi.org/project/vid2cc-ai/) [![License: MIT](https://img.shields.io/badge/License-MIT-blue.svg)](https://opensource.org/licenses/MIT) [![Python 3.9+](https://img.shields.io/badge/python-3.9+-green.svg)](https://www.python.org/downloads/) [![Code Style: Black](https://img.shields.io/badge/code%20style-black-000000.svg)](https://github.com/psf/black) ![CI Status](https://github.com/0xdilshan/vid2cc-AI/actions/workflows/ci.yml/badge.svg) **vid2cc-AI** is a high-performance CLI tool designed to bridge the gap between raw video and accessible content. By leveraging OpenAI's Whisper models and FFmpeg's robust media handling, it automates the creation of perfectly synced `.srt` subtitles. --- ## Table of contents - [🚀 Key Features](#-key-features) - [⚙️ Installation](#-installation) - [1. Prerequisite: FFmpeg](#1-prerequisite-ffmpeg) - [2. Install vid2cc-AI](#2-install-vid2cc-ai) - [📖 How To Use](#-how-to-use) - [🛠️ Advanced Options](#-advanced-options) - [📦 Batch Processing](#-batch-processing) - [📦 Usage as a Library](#-usage-as-a-library) - [☁️ Run on Google Colab (with UI)](#run-on-google-colab-with-ui) - [🧪 Testing](#-testing) - [🗺️ Roadmap](#-roadmap) - [🛠️ Tech Stack](#-tech-stack) - [📄 License](#-license) --- ## 🚀 Key Features - **AI-Driven Transcription:** Powered by OpenAI Whisper for industry-leading accuracy. - **Hardware Acceleration:** Automatic CUDA detection for GPU-accelerated processing. - **Intelligent Pre-processing:** FFmpeg-based audio extraction optimized for speech recognition (16kHz Mono). - **Professional Packaging:** Fully installable via pip with a dedicated command-line entry point. --- ## ⚙️ Installation ### 1. Prerequisite: FFmpeg This tool requires **FFmpeg** to be installed on your system. For a complete step-by-step guide on how to install FFmpeg on Windows (Winget/Choco), macOS (Homebrew), or Linux (Apt/Dnf/Pacman), please refer to the dedicated guide: 👉 **[FFmpeg Installation Guide](./ffmpeg_installation.md)** ### 2. Install vid2cc-AI ```bash pip install vid2cc-ai ``` **Install directly from the source for development:* ```bash git clone https://github.com/0xdilshan/vid2cc-AI.git cd vid2cc-AI pip install -e . ``` ## 📖 How To Use Once installed, the `vid2cc` command is available globally in your terminal. #### Examples *For maximum accuracy with toggleable subs:* ```bash vid2cc example.mp4 --model large --embed ``` --- ### 🛠️ Advanced Options Fine-tune your output using the following flags: | Flag | Description | | :--- | :--- | | `--model [size]` | **Choose Whisper model:** `tiny`, `base`, `small`, `medium`, `large` or `turbo`. | | `--embed` | **Soft Subtitles:** Adds the SRT as a metadata track. Fast and allows users to toggle subtitles on/off in players like VLC. | | `--hardcode` | **Burn-in Subtitles:** Permanently draws subtitles onto the video. Essential for social media (Instagram/TikTok) where players don't support SRT files. | | `--output-dir` or `-o` | **Set Output Directory:** Create the destination directory if it doesn't exist and ensure all generated files (SRT, audio, and video) are saved there. | | `--translate` or `-t` | **Translate to English:** Automatically translate any supported language transcription to English | ### 📦 Batch Processing No need to run the command for every single file. You can pass multiple videos at once: ```bash # Process all mp4 files in the current directory vid2cc *.mp4 --model small --embed # Process multiple specific files vid2cc video1.mp4 video2.mkv video3.mov --model base --embed ``` --- ### 📦 Usage as a Library You can integrate **vid2cc-AI** directly into your Python projects: ```python from vid2cc_ai import Transcriber, extract_audio # Extract and Transcribe extract_audio("video.mp4", "audio.wav") ts = Transcriber("base") segments = ts.transcribe("audio.wav") for s in segments: print(f"[{s['start']:.2f}s] {s['text']}") ``` --- ## Run on Google Colab (with UI) You can run vid2cc-ai directly in your browser using Google Colab. This version includes a friendly interface to manage your Google Drive files and transcription settings without writing code. 1. **Open the Notebook:** [![Open in Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/0xdilshan/vid2cc-AI/blob/main/Vid2CC_AI.ipynb). 2. **Install & Mount:** Run the first cell to install `vid2cc-ai` and connect your Google Drive. 3. **Configure UI:** * **Video Path:** Right-click your video in the Colab file sidebar and select "Copy Path." - **Model:** Choose `turbo` or `small` for speed, large for accuracy. - **Output:** Select if you want Soft Subtitles (toggleable) or Hardcoded (burned-in). 4. **Start:** Click **"Start Processing"** and find your result in your Drive folder. ⚡ For 10x faster transcription, ensure your Colab runtime is set to GPU (`Runtime` > `Change runtime type` > `T4 GPU`). [![colab-notebook-preview](img/google-colab-example.jpg)](https://colab.research.google.com/github/0xdilshan/vid2cc-AI/blob/main/Vid2CC_AI.ipynb) --- ## 🧪 Testing ```bash # Install test dependencies pip install pytest # Run the test suite pytest ``` --- ## 🗺️ Roadmap - [x] Local video → SRT subtitle/ transcription - [x] Embed subtitles into video containers (`--embed`) - [x] Burn-in subtitles (`--hardcode`) - [x] Set custom output directory (`--output-dir`) - [x] Multilingual transcription - [x] Support translation to English - ~~[ ] Transcription from YouTube/Vimeo URLs (`yt-dlp`)~~ - [x] Google Colab notebook support --- ## 🛠️ Tech Stack - **Inference:** OpenAI Whisper - **Media Engine:** FFmpeg - **Core:** Python 3.9+, PyTorch - **CLI Framework:** Argparse --- ## 📄 License Distributed under the MIT License. See `LICENSE` for more information.
text/markdown
Dilshan
null
null
null
null
ffmpeg, srt, cc, whisper-ai, subtitle, whisper
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "Topic :: Multimedia :: Video", "Topic :: Multimedia :: Sound/Audio :: Speech", "Topic :: Scientific/Engineering :: Artificial Intelligence", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3" ]
[]
null
null
>=3.9
[]
[]
[]
[ "openai-whisper", "torch", "tqdm" ]
[]
[]
[]
[ "Homepage, https://github.com/0xdilshan/vid2cc-AI", "Source, https://github.com/0xdilshan/vid2cc-AI", "Documentation, https://github.com/0xdilshan/vid2cc-AI#readme", "Issues, https://github.com/0xdilshan/vid2cc-AI/issues", "Releases, https://github.com/0xdilshan/vid2cc-AI/releases" ]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:08:45.172580
vid2cc_ai-0.1.5.tar.gz
11,906
66/11/0a834961b3b664f1a7160f3d58223d74b4fd61d7eadc934ece5af4ef89fe/vid2cc_ai-0.1.5.tar.gz
source
sdist
null
false
e7b6a091cab6e0d564140863d7dd298a
e818d1d562d8fb88c35d8ed2a7cbb5ca4076baf11e8c322a4a30845ebfd00da0
66110a834961b3b664f1a7160f3d58223d74b4fd61d7eadc934ece5af4ef89fe
null
[ "LICENSE" ]
254
2.2
turbx
1.1.6
Extensible toolkit for analyzing turbulent flow datasets
# turbx [![PyPI version](https://badge.fury.io/py/turbx.svg)](https://badge.fury.io/py/turbx) [![Downloads](https://pepy.tech/badge/turbx)](https://pepy.tech/project/turbx) <!--- [![pipeline status](https://gitlab.iag.uni-stuttgart.de/transi/turbx/badges/master/pipeline.svg)](https://gitlab.iag.uni-stuttgart.de/transi/turbx/-/commits/master) --> ## About `turbx` is a package for processing turbulent flow datasets. Primary data access classes are `super()`ed wrappers of `h5py.File` that make data & metadata access tidy and performant. Workloads requiring heavy I/O and compute are made scalable using parallelization and high-performance collective MPI-IO. ## Pre-installation requirements ### `h5py` `turbx` is designed to be used with parallel-mode `h5py`, the `python3` API for `HDF5`. Most basic workflows will still work transparently with the basic serial install of `h5py`, however data-heavy workflows will require the full API functionality. This requires: - A parallel `HDF5` installation - `h5py` built with `HDF5_MPI="ON"`; see [h5py docs](https://docs.h5py.org/en/stable/mpi.html) <br> Confirm that `h5py` was built with MPI support: ```python >>> import h5py >>> h5py.get_config().mpi True ``` ### `mpi4py` High-performance collective MPI-IO and MPI operations are handled with `mpi4py`. This requires: - An MPI implementation such as `OpenMPI` or `MPICH` - `mpi4py`; see [mpi4py docs](https://mpi4py.readthedocs.io/en/stable/install.html) ## Installation ### TL;DR Install binary directly from [PyPI](https://pypi.org/project/turbx): ``` python3 -m pip install --upgrade turbx ``` --- ### Non-root install with `--user` The `--user` flag can be added to install to `~/.local/lib/pythonX.Y/site-packages` rather than `site-packages` of the `python3` installation itself. This is often required for HPC environments where installing packages for the system `python3` is not allowed for regular users. <br> ### Editable installs `turbx` can also be installed from source in `editable` mode (see [setuptools docs](https://setuptools.pypa.io/en/latest/userguide/development_mode.html)). Once the source is acquired (PyPI or GitLab), the source can be installed from the project root folder: ``` python3 -m pip install --upgrade -e . ``` <br> ### Installing on systems with no outbound network access If the restricted install environment requires `devpi-server`: **On a local machine with internet access:** ``` devpi-server ``` **On a local machine with internet access:** ``` ssh -R <remote_port>:localhost:<local_devpi_port> user@domain.com ``` - `<local_devpi_port>` is the port on which `devpi-server` is running locally - `<remote_port>` is an arbitrary free port on the system login node used to expose the mirror remotely - After connecting, the PyPI mirror will be reachable on the server side at: `http://localhost:<remote_port>` **On the remote system:** ``` python3 -m pip install --upgrade [--user] [--editable] --index-url http://localhost:<remote_port>/root/pypi/+simple/ <package-or-path> ```
text/markdown
Jason A
null
Jason A
null
MIT
turbulence, fluid mechanics, DNS, MPI, HPC
[ "License :: OSI Approved :: MIT License", "Development Status :: 4 - Beta", "Intended Audience :: Science/Research", "Programming Language :: Python :: 3 :: Only", "Topic :: Scientific/Engineering :: Physics", "Topic :: System :: Distributed Computing", "Topic :: Database" ]
[]
null
null
>=3.11
[]
[]
[]
[ "mpi4py>=4.0", "numpy>=2.0", "scipy>=1.14", "h5py>=3.11", "matplotlib>=3.9", "tqdm>=4.66", "psutil>=6.0", "ruff>=0.6; extra == \"dev\"", "pytest>=8.0; extra == \"dev\"", "pyright>=1.1; extra == \"dev\"" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.12.3
2026-02-19T12:08:25.677674
turbx-1.1.6.tar.gz
207,845
58/30/65ba85d5b71dd55cb68f52850cf1cc62682f1cb6f8bf562e4d54aff9c505/turbx-1.1.6.tar.gz
source
sdist
null
false
99a52453a530821ed7d8e798d501d158
08ea10d41bb49ee78e871d313061c9d867921fe1d5e914434147da6d1a003509
583065ba85d5b71dd55cb68f52850cf1cc62682f1cb6f8bf562e4d54aff9c505
null
[]
237
2.1
ipi
3.1.10
A Python interface for ab initio path integral molecular dynamics simulations
# i-PI: a Universal Force Engine A Python interface for ab initio path integral molecular dynamics simulations (and more). i-PI is a Python server (that does not need to be compiled and only requires a relatively recent version of Python and Numpy) that applies an algorithm to update the positions of the nuclei. One of many compatible external codes acts as client, and computes the electronic energy and forces. This is typically a patched version of an electronic structure code, but a simple self-contained Fortran driver that implements several simple interatomic potentials is included for test purposes. i-PI was originally developed to simulate the quantum mechanical nature of light nuclei by performing path integral molecular dynamics simulations, and it implements most of the state-of-the-art methods to accelerate this kind of calculations. It has since grown to also provide all sorts of simulation strategies, from replica exchange to geometry optimization. If you use i-PI in your research, please cite the accompanying publication: for version 3, the relevant paper is [Litman et al., _J. Chem. Phys._ 161, 062504 (2024)](https://doi.org/10.1063/5.0215869) ``` @article{litman2024ipi, title={i-PI 3.0: a flexible and efficient framework for advanced atomistic simulations}, author={Yair Litman and Venkat Kapil and Yotam M. Y. Feldman and Davide Tisi and Tomislav Begušić and Karen Fidanyan and Guillaume Fraux and Jacob Higer and Matthias Kellner and Tao E. Li and Eszter S. Pós and Elia Stocco and George Trenins and Barak Hirshberg and Mariana Rossi and Michele Ceriotti}, journal = {J. Chem. Phys.}, pages = {062505}, volume = {161}, year = {2024} } ``` ## Quick Setup To use i-PI with an existing driver, install and update using `pip`: Last version: ```bash python -m pip install git+https://github.com/i-pi/i-pi.git ``` Last Release: ```bash pip install -U ipi ``` ## Documentation You can find the online documentation at [https://docs.ipi-code.org](https://docs.ipi-code.org/). Alternatively, you can build it locally by following instructions in the `docs/README.md` file. ## Source installation To develop i-PI or test it with the self-contained driver, follow these instructions. It is assumed that i-PI will be run from a Linux environment, with a recent version of Python, Numpy and gfortran, and that the terminal is initially in the i-pi package directory (the directory containing this file), which you can obtain by cloning the repository ```bash git clone https://github.com/i-pi/i-pi.git ``` Source the environment settings file `env.sh` as `source env.sh` or `. env.sh`. It is useful to put this in your `.bashrc` or other settings file if you always want to have i-PI available. ## Compile the driver code The built-in driver requires a FORTRAN compiler, and can be built as ```bash cd drivers/f90 make cd ../.. ``` There is also a Python driver available in `drivers/py`, which however has limited functionalities. ## Examples and demos The `examples` and `demos` folders contain inputs for many different types of calculations based on i-PI. Examples are typically minimal use-cases of specific features, while demos are more structured, tutorial-like examples that show how to realize more complex setups, and also provide a brief discussion of the underlying algorithms. To run these examples, you should typically start i-PI, redirecting the output to a log file, and then run a couple of instances of the driver code. The progress of the wrapper is followed by monitoring the log file with the `tail` Linux command. Optionally, you can make a copy of the directory with the example somewhere else if you want to keep the i-PI directory clean. For example, after sourcing the `env.sh` file, ```bash cd demos/para-h2-tutorial/tutorial-1/ i-pi tutorial-1.xml > log & i-pi-driver -a localhost -p 31415 -m sg -o 15 & i-pi-driver -a localhost -p 31415 -m sg -o 15 & tail -f log ``` The monitoring can be interrupted with CTRL+C when the run has finished (5000 steps). ## Tutorials and online resources The i-PI [documentation](https://docs.ipi-code.org/onlinereso.html) has a list of available tutorials, recipes and other useful online resources. ## Run the automatic test suite The automatic test suite can be run by calling the i-pi-tests script. You need to have the `pytest` package installed ``` i-pi-tests ``` You may also need to install some dependencies, listed in `requirements.txt`. See more details in the README file inside the `ipi_tests` folder. ## Contributing If you have new features you want to implement into i-PI, your contributions are much welcome. See `CONTRIBUTING.md` for a brief set of style guidelines and best practices. Before embarking into a substantial project, it might be good to get in touch with the developers, e.g. by opening a wishlist issue.
text/x-rst
The i-PI developers
ipi.managers@gmail.com
null
null
null
null
[ "Development Status :: 5 - Production/Stable", "Environment :: Console", "Intended Audience :: Science/Research", "Operating System :: MacOS :: MacOS X", "Operating System :: POSIX :: Linux", "License :: OSI Approved :: GNU General Public License v3 (GPLv3)", "Programming Language :: Python :: 3", "Pr...
[]
http://ipi-code.org
null
>=3.8
[]
[]
[]
[ "numpy" ]
[]
[]
[]
[ "Documentation, https://ipi-code.org/i-pi", "Repository, https://github.com/i-pi/i-pi" ]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:08:12.163796
ipi-3.1.10.tar.gz
16,917,908
f1/bf/b2e5ff4ec1d3f17411c56a33102c3d9ac99f1e824b4e6866de74d00c29e7/ipi-3.1.10.tar.gz
source
sdist
null
false
bb65d77865384b7e12f04d1e166a4b26
bfe818c2b66f5392c62ae449605b9e76ef61df184b0a5b05c96752e648f8bcdb
f1bfb2e5ff4ec1d3f17411c56a33102c3d9ac99f1e824b4e6866de74d00c29e7
null
[]
1,272
2.4
azure-kusto-ingest
6.0.2
Kusto Ingest Client
Microsoft Azure Kusto Ingest Library for Python =============================================== .. code-block:: python from azure.kusto.data import KustoConnectionStringBuilder, DataFormat from azure.kusto.ingest import QueuedIngestClient, IngestionProperties, FileDescriptor, BlobDescriptor ingestion_props = IngestionProperties(database="{database_name}", table="{table_name}", data_format=DataFormat.CSV) client = QueuedIngestClient(KustoConnectionStringBuilder.with_interactive_login("https://ingest-{cluster_name}.kusto.windows.net")) file_descriptor = FileDescriptor("{filename}.csv", 15360) # in this example, the raw (uncompressed) size of the data is 15KB (15360 bytes) client.ingest_from_file(file_descriptor, ingestion_properties=ingestion_props) client.ingest_from_file("{filename}.csv", ingestion_properties=ingestion_props) blob_descriptor = BlobDescriptor("https://{path_to_blob}.csv.gz?sas", 51200) # in this example, the raw (uncompressed) size of the data is 50KB (52100 bytes) client.ingest_from_blob(blob_descriptor, ingestion_properties=ingestion_props) Overview -------- *Kusto Python Ingest Client* Library provides the capability to ingest data into Kusto clusters using Python. It is Python 3.x compatible and supports data types through familiar Python DB API interface. It's possible to use the library, for instance, from `Jupyter Notebooks <http://jupyter.org/>`_ which are attached to Spark clusters, including, but not exclusively, `Azure Databricks <https://azure.microsoft.com/en-us/services/databricks>`_ instances. * `How to install the package <https://github.com/Azure/azure-kusto-python#install>`_. * `Data ingest sample <https://github.com/Azure/azure-kusto-python/blob/master/azure-kusto-ingest/tests/sample.py>`_. * `GitHub Repository <https://github.com/Azure/azure-kusto-python/tree/master/azure-kusto-data>`_.
text/x-rst
Microsoft Corporation
Microsoft Corporation <kustalk@microsoft.com>
null
null
null
azure, kusto, ingest, data, analytics
[ "Programming Language :: Python", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "License :: OSI Approved :: MIT License" ]
[]
null
null
>=3.9
[]
[]
[]
[ "azure-kusto-data", "azure-storage-blob==12.26.0", "azure-storage-queue==12.13.0", "tenacity>=9.1.2", "pandas>=2.3.1; extra == \"pandas\"" ]
[]
[]
[]
[ "Repository, https://github.com/Azure/azure-kusto-python" ]
twine/6.1.0 CPython/3.12.8
2026-02-19T12:06:38.719256
azure_kusto_ingest-6.0.2.tar.gz
19,824
d5/33/e4511ff217ef72ff4d286fcbc1633fc89c5b5b5c9eae447b21b4b1cbe2ff/azure_kusto_ingest-6.0.2.tar.gz
source
sdist
null
false
dae4ff27a766e264bab4f0db82182820
ebc1570de9f34b3ce075b332d4f1205631575f5a524c7bdc84f81e5f33157d1a
d533e4511ff217ef72ff4d286fcbc1633fc89c5b5b5c9eae447b21b4b1cbe2ff
MIT
[]
199,245
2.4
azure-kusto-data
6.0.2
Kusto Data Client
Microsoft Azure Kusto Library for Python ======================================== Overview -------- .. code-block:: python from azure.kusto.data import KustoClient, KustoConnectionStringBuilder cluster = "<insert here your cluster name>" client_id = "<insert here your AAD application id>" client_secret = "<insert here your AAD application key>" authority_id = "<insert here your AAD tenant id>" kcsb = KustoConnectionStringBuilder.with_aad_application_key_authentication(cluster, client_id, client_secret, authority_id) # It is a good practice to re-use the KustoClient instance, as it maintains a pool of connections to the Kusto service. # This sample shows how to create a client and close it in the same scope, for demonstration purposes. with KustoClient(kcsb) as client: db = "Samples" query = "StormEvents | take 10" response = client.execute(db, query) for row in response.primary_results[0]: print(row[0], " ", row["EventType"]) *Kusto Python Client* Library provides the capability to query Kusto clusters using Python. It is Python 3.x compatible and supports all data types through familiar Python DB API interface. It's possible to use the library, for instance, from `Jupyter Notebooks <http://jupyter.org/>`_. which are attached to Spark clusters, including, but not exclusively, `Azure Databricks <https://azure.microsoft.com/en-us/services/databricks/>`_. instances. Async Client ~~~~~~~~~~~~ Kusto now provides an asynchronous client for queries. To use the client, first install the package with the "aio" extra: .. code:: bash pip install azure-kusto-data[aio] The async client uses exact same interface as the regular client, except that it lives in the ``azure.kusto.data.aio`` namespace, and it returns ``Futures`` you will need to ``await`` its .. code:: python from azure.kusto.data import KustoConnectionStringBuilder from azure.kusto.data.aio import KustoClient cluster = "<insert here your cluster name>" client_id = "<insert here your AAD application id>" client_secret = "<insert here your AAD application key>" authority_id = "<insert here your AAD tenant id>" async def sample(): kcsb = KustoConnectionStringBuilder.with_aad_application_key_authentication(cluster, client_id, client_secret, authority_id) async with KustoClient(kcsb) as client: db = "Samples" query = "StormEvents | take 10" response = await client.execute(db, query) for row in response.primary_results[0]: print(row[0], " ", row["EventType"]) Links ~~~~~ * `How to install the package <https://github.com/Azure/azure-kusto-python#install>`_. * `Kusto query sample <https://github.com/Azure/azure-kusto-python/blob/master/azure-kusto-data/tests/sample.py>`_. * `GitHub Repository <https://github.com/Azure/azure-kusto-python/tree/master/azure-kusto-data>`_.
text/x-rst
Microsoft Corporation
Microsoft Corporation <kustalk@microsoft.com>
null
null
null
azure, kusto, data, analytics
[ "Programming Language :: Python", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "License :: OSI Approved :: MIT License" ]
[]
null
null
>=3.9
[]
[]
[]
[ "python-dateutil>=2.9.0", "requests>=2.32.4", "azure-identity<2,>=1.24.0", "msal<2,>=1.34.0b1", "ijson~=3.4.0", "azure-core<2,>=1.35.0", "aiohttp>=3.12.15; extra == \"aio\"", "asgiref>=3.9.1; extra == \"aio\"", "pandas>=2.3.1; extra == \"pandas\"" ]
[]
[]
[]
[ "Repository, https://github.com/Azure/azure-kusto-python" ]
twine/6.1.0 CPython/3.12.8
2026-02-19T12:06:37.867769
azure_kusto_data-6.0.2.tar.gz
39,679
96/d7/02b9286bfc1fcd3ab16c5f02ad89be4e3747915ce9b274e96d45141104d0/azure_kusto_data-6.0.2.tar.gz
source
sdist
null
false
eeb72b9f5d356d20c23e44086e5a3134
e99ca60e9518a8bd41e947b53f5bb727ec2f08ce6f1c58cba200db5b9ad6f3ce
96d702b9286bfc1fcd3ab16c5f02ad89be4e3747915ce9b274e96d45141104d0
MIT
[]
441,975
2.4
ai-rulez
3.7.3
⚡ One config to rule them all. Centralized AI assistant configuration management - generate rules for Claude, Cursor, Copilot, Windsurf and more from a single YAML file.
# ai-rulez Write AI assistant rules once. Generate configs for 18 different tools. Every AI coding tool wants its own config format. Claude needs `CLAUDE.md`, Cursor wants `.cursor/rules/`, Copilot expects `.github/copilot-instructions.md`. Keeping them in sync is tedious and error-prone. ai-rulez solves this: define rules and context in `.ai-rulez/`, run `generate`, and get native configs for all your tools. ```bash npx ai-rulez@latest init && npx ai-rulez@latest generate ``` **[Documentation](https://goldziher.github.io/ai-rulez/)** ## What You Get - **18 preset generators**: Claude, Cursor, Windsurf, Copilot, Gemini, Cline, Continue.dev, Amp, Junie, Codex, OpenCode, and custom presets - **Commands system**: Define slash commands once, use them across tools that support it - **Context compression**: 34% size reduction with smart whitespace optimization - **Remote includes**: Pull shared rules from git repos (company standards, team configs) - **Profile system**: Generate different configs for backend/frontend/QA teams - **MCP server**: Let AI assistants manage their own rules via Model Context Protocol - **Type-safe schemas**: JSON Schema validation for all config files ## Quick Start ```bash # No install required npx ai-rulez@latest init "My Project" npx ai-rulez@latest generate ``` This creates: ``` .ai-rulez/ ├── config.yaml # Which tools to generate for ├── rules/ # Guidelines AI must follow ├── context/ # Project background info ├── skills/ # Specialized AI roles ├── agents/ # Agent-specific prompts └── commands/ # Slash commands ``` And generates native configs for each tool you specify. ## Configuration ```yaml # .ai-rulez/config.yaml version: "3.0" name: "My Project" presets: - claude - cursor - copilot - windsurf # Optional: team-specific profiles profiles: backend: [backend, database] frontend: [frontend, ui] # Optional: share rules across repos includes: - name: company-standards source: https://github.com/company/ai-rules.git ref: main ``` ## Content Structure **Rules** - What AI must do: ```markdown --- priority: critical --- # Security Standards - Never commit credentials - Use environment variables for secrets - Sanitize all user input ``` **Context** - What AI should know: ```markdown --- priority: high --- # Architecture This is a microservices app: - API Gateway (Go, port 8080) - Auth Service (Go, port 8081) - PostgreSQL 15 ``` **Commands** - Slash commands across tools: ```markdown --- name: review aliases: [r, pr-review] targets: [claude, cursor, continue-dev] --- # Code Review Review the current PR for: 1. Logic errors 2. Security issues 3. Performance problems ``` ## Installation **No install required:** ```bash npx ai-rulez@latest <command> # or uvx ai-rulez <command> ``` **Global install:** ```bash # Homebrew brew install goldziher/tap/ai-rulez # npm npm install -g ai-rulez # pip pip install ai-rulez # Go go install github.com/Goldziher/ai-rulez/cmd@latest ``` ## CLI Reference ```bash # Initialize project ai-rulez init "Project Name" ai-rulez init --domains backend,frontend,qa # Generate configs ai-rulez generate ai-rulez generate --profile backend ai-rulez generate --dry-run # Content management ai-rulez add rule security-standards --priority critical ai-rulez add context api-docs ai-rulez add skill database-expert ai-rulez add command review-pr ai-rulez list rules ai-rulez remove rule outdated-rule # Validation ai-rulez validate # MCP server (for AI assistants) npx ai-rulez@latest mcp # Migrate from V2 ai-rulez migrate v3 ``` ## Remote Includes Share rules across repositories: ```yaml includes: # HTTPS - name: company-standards source: https://github.com/company/ai-rules.git ref: main include: [rules, context] merge_strategy: local-override # SSH - name: shared-configs source: git@github.com:org/shared-ai-rulez.git ref: v2.0.0 include: [rules, skills] # Local path - name: local-standards source: ../shared-rules include: [rules] ``` Private repos use `AI_RULEZ_GIT_TOKEN` environment variable or `--token` flag. ## Generated Output Running `ai-rulez generate` creates: | Preset | Output | |--------|--------| | Claude | `CLAUDE.md` + `.claude/skills/` + `.claude/agents/` | | Cursor | `.cursor/rules/*.mdc` | | Windsurf | `.windsurf/*.md` | | Copilot | `.github/copilot-instructions.md` | | Gemini | `.gemini/config.yaml` | | Continue.dev | `.continue/prompts/ai_rulez_prompts.yaml` | | Cline | `.cline/rules/*.md` | | Custom | Any path with markdown, JSON, or directory output | ## Use Cases **Monorepo**: Generate configs for multiple packages ```bash ai-rulez generate --recursive ``` **Team profiles**: Different rules for different teams ```bash ai-rulez generate --profile backend ai-rulez generate --profile frontend ``` **CI validation**: Ensure configs stay in sync ```bash ai-rulez validate && ai-rulez generate --dry-run ``` **Import existing configs**: Migrate from tool-specific files ```bash ai-rulez init --from auto ai-rulez init --from .cursorrules,CLAUDE.md ``` ## MCP Server Let AI assistants manage rules directly: ```yaml # .ai-rulez/mcp.yaml version: "3.0" mcp_servers: - name: ai-rulez command: npx args: ["-y", "ai-rulez@latest", "mcp"] transport: stdio enabled: true ``` The MCP server exposes CRUD operations, validation, and generation to AI assistants. ## Compression Reduce context size for token-constrained tools: ```yaml compression: level: standard # none, minimal, standard, aggressive ``` At `standard` level, output is ~34% smaller through whitespace optimization and duplicate removal. ## Documentation - [Configuration Reference](https://goldziher.github.io/ai-rulez/configuration/) - [Domains & Profiles](https://goldziher.github.io/ai-rulez/domains/) - [Remote Includes](https://goldziher.github.io/ai-rulez/includes/) - [MCP Server](https://goldziher.github.io/ai-rulez/mcp-server/) - [Schema Validation](https://goldziher.github.io/ai-rulez/schema/) - [Migration Guide](https://goldziher.github.io/ai-rulez/migration/) ## Contributing Contributions welcome. See [CONTRIBUTING.md](https://github.com/Goldziher/ai-rulez/blob/main/CONTRIBUTING.md). ## License MIT
text/markdown
Na'aman Hirschfeld
nhirschfeld@gmail.com
null
null
null
ai, ai-assistant, ai-rules, claude, cursor, copilot, windsurf, gemini, cline, continue-dev, mcp, model-context-protocol, cli, configuration, config, rules, generator, golang, go, development, developer-tools, automation, workflow, productivity, pre-commit, git-hooks, code-generation, ai-development, assistant-configuration, monorepo, presets, agents
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "Topic :: Software Development :: Code Generators", "Topic :: Software Development :: Libraries :: Python Modules", "Topic :: Software Development :: Quality Assurance", "Topic :: Software Development :: Pre-processors", "Topic :: Text...
[]
https://github.com/Goldziher/ai-rulez
null
>=3.8
[]
[]
[]
[]
[]
[]
[]
[ "Homepage, https://goldziher.github.io/ai-rulez/", "Documentation, https://goldziher.github.io/ai-rulez/", "Bug Reports, https://github.com/Goldziher/ai-rulez/issues", "Source, https://github.com/Goldziher/ai-rulez", "Changelog, https://github.com/Goldziher/ai-rulez/releases", "Funding, https://github.com...
twine/6.1.0 CPython/3.13.7
2026-02-19T12:06:36.466778
ai_rulez-3.7.3.tar.gz
11,057
1c/82/ad794b61b13bb9154ab6a01b142073ebbdca9cd6b4fdafd10141174c3575/ai_rulez-3.7.3.tar.gz
source
sdist
null
false
80c087bf656fcf17e74b4d173a6c4c06
81ee7f0b4f649924dd0bb7d6a0795e9331e09f834de65dfdb9e86c0100b9174f
1c82ad794b61b13bb9154ab6a01b142073ebbdca9cd6b4fdafd10141174c3575
null
[]
255
2.4
vedana-core
0.7.3
Semantic Graph RAG App
# Vedana
text/markdown
null
Andrey Tatarinov <a@tatarinov.co>, Timur Sheydaev <tsheyd@epoch8.co>
null
null
null
null
[]
[]
null
null
>=3.12
[]
[]
[]
[ "aiohttp>=3.11.18", "aioitertools>=0.12.0", "aiosqlite>=0.21.0", "alembic>=1.13.0", "async-lru>=2.0.5", "asyncpg>=0.29.0", "grist-api>=0.1.1", "jims-core>=0.5.1", "jims-telegram>=0.5.1", "litellm>=1.79.0", "neo4j>=5.28.1", "openai>=2.8.0", "pandas>=2.2.3", "psycopg2-binary>=2.9.10", "pyd...
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:06:32.965750
vedana_core-0.7.3.tar.gz
30,497
a5/d3/1db53d9304f103fea46dc12b9fa089c95e1cada18bc81e6a223ad1beac86/vedana_core-0.7.3.tar.gz
source
sdist
null
false
edc94cd5ca2936507524a09e1326c988
9ea9bb15137680f6cf3f59a48d6a43fb4f7c58057e61857afc20b45ab2a05ef7
a5d31db53d9304f103fea46dc12b9fa089c95e1cada18bc81e6a223ad1beac86
null
[]
250
2.4
jims-api
0.5.2
universal API for JIMS application
# jims-api `jims-api` exposes any JIMS application as a FastAPI HTTP service. ## Run ```bash jims-api --app my_project.app:app --port 8080 ``` Environment variables: - `JIMS_APP` - JIMS app import path (`module:attr`) - `JIMS_PORT` - HTTP port - `JIMS_HOST` - HTTP host - `JIMS_API_KEY` - optional bearer token for auth ## Endpoints - `GET /health` - `POST /api/v1/chat` ### `POST /api/v1/chat` Request: ```json { "contact_id": "customer:42", "message": "Hello", "thread_id": null, "thread_config": {"interface": "api"}, "event_type": "comm.user_message", "run_conversation_start_on_new_thread": false } ``` Response: ```json { "thread_id": "0194f0f3-d88d-7cca-8f37-ff44f911f539", "created_new_thread": true, "assistant_messages": ["Hi! How can I help?"], "events": [ { "event_type": "comm.assistant_message", "event_data": {"role": "assistant", "content": "Hi! How can I help?"} } ] } ``` If `JIMS_API_KEY` is set, send it as `Authorization: Bearer <token>`
text/markdown
null
null
null
null
null
null
[]
[]
null
null
>=3.12
[]
[]
[]
[ "click>=8.0", "fastapi>=0.128.0", "jims-core>=0.5.1", "loguru>=0.7.3", "uvicorn>=0.35.0" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:06:31.745127
jims_api-0.5.2.tar.gz
7,608
ae/0e/c1a1f20d654dfc497c503556ac92b677bc890ad2148f407225597435f4f4/jims_api-0.5.2.tar.gz
source
sdist
null
false
4a897376b36d767ef1bd256b06e92996
50196db683ab20defea2ecc2ecc269ee33eed92f7bc08b4e42c0e7838c8ca2e2
ae0ec1a1f20d654dfc497c503556ac92b677bc890ad2148f407225597435f4f4
null
[]
240
2.4
jims-core
0.5.2
Add your description here
# Name JIMS stands for "Just an Integrated Multiagent System" # Concepts ## Thread Thread is a conversation between user (or users) and agentic system Each interaction between agents and users happens in thread ## Event Something that happens in a thread Each event has id, type and data Example: { "event_id": "...", "event_type": "comm.user_message", "event_data": { "content": "Hello!" } }
text/markdown
null
Andrey Tatarinov <a@tatarinov.co>
null
null
null
null
[]
[]
null
null
>=3.12
[]
[]
[]
[ "backoff>=2.2.1", "litellm>=1.79.0", "openinference-instrumentation-litellm>=0.1.25", "opentelemetry-api>=1.34.1", "opentelemetry-exporter-otlp", "opentelemetry-sdk>=1.34.1", "prometheus-client>=0.21.1", "pydantic-settings>=2.8.1", "sqlalchemy[asyncio]>=2.0.40", "uuid7>=0.1.0" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:06:29.659321
jims_core-0.5.2.tar.gz
14,419
46/aa/e69027bc6208a6ed47919153686e89902c0f43ceee8edda9a884b1797851/jims_core-0.5.2.tar.gz
source
sdist
null
false
7d9084a9de970d499508469840d20679
19dad443719918fb457fe4290cf126a8ff02bb9924b6dbbbf9d57c74783169ec
46aae69027bc6208a6ed47919153686e89902c0f43ceee8edda9a884b1797851
null
[]
263
2.4
vedana-etl
0.7.3
Pipeline template for Vedana
# Basic pipeline for all vedana projects. This pipeline: - Parses Grist Data & Data Model - Ensures that Memgraph index/vector index structure is in sync with data model - Updates Memgraph database in incremental fashion To add steps: 1. Pass extra transformations to [get_pipeline](src/pipeline.py) 2. Create new app configuration from [app.py](src/app.py) ## Pipeline Labels Hierarchy ### Pipeline `labels=("pipeline", "pipeline_name")` defines a set of operations as standalone, sort of like a DAG in Airflow or a Dagster Job. Its purpose is to be able to render it as a separate tab on the ETL page of Backoffice in order to look at it independently of other transformations ### Stage `labels=("stage", "stage_name")` defines a stage of `pipeline`. Currently, stages are useful for creating and managing observability features, such as [main dashboard's](/libs/vedana-backoffice/vedana_backoffice/pages/main_dashboard.py) Ingest table, which displays DataTable's of all transformations with `labels=("stage", "extract")`. Stages are also useful when running the pipeline manually. ### Flow `labels=("flow", "flow_name")` helps execute a `pipeline` (or possibly several pipelines) in a nice fashion, used in defining cron jobs, etc.
text/markdown
null
Andrey Tatarinov <a@tatarinov.co>, Timur Sheydaev <tsheyd@epoch8.co>
null
null
null
null
[]
[]
null
null
>=3.12
[]
[]
[]
[ "alembic>=1.16.1", "datapipe-app>=0.5.4", "datapipe-core>=0.14.8", "grist-api>=0.1.1", "neo4j>=5.28.1", "openai>=2.8.0", "pandas>=1.2.0", "pgvector>=0.4.2", "pytest>=8.4.1", "requests>=2.32.4", "sqlalchemy>=2.0.41", "vedana-core>=0.6.3" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:06:26.358913
vedana_etl-0.7.3.tar.gz
28,873
72/c6/1c8c3a51095890c049fac63e863678a1edbcc0d50f4ef1fdfba18422f51b/vedana_etl-0.7.3.tar.gz
source
sdist
null
false
ff04ad9030f5bc7bc880d20f5d8624f4
761d242557a8a1fb0df4e28f36225afdf78b020031d59d26fb48febdf80cea80
72c61c8c3a51095890c049fac63e863678a1edbcc0d50f4ef1fdfba18422f51b
null
[]
251
2.4
jims-tui
0.5.2
Add your description here
Simple TUI for interactive demo of ThreadController and given answering pipeline ## Run as TUI ``` jims-tui --app jims_demo.app ``` ## Run as Telegram bot ``` jims-telegram --app jims_demo.app ```
text/markdown
null
Andrey Tatarinov <a@tatarinov.co>
null
null
null
null
[]
[]
null
null
>=3.12
[]
[]
[]
[ "textual>=3.1.0" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:06:25.093182
jims_tui-0.5.2.tar.gz
6,187
3f/96/366c4ae531effd3f4266b231c796ba5d6d93827dab244190c49d30631ea1/jims_tui-0.5.2.tar.gz
source
sdist
null
false
5da19ec0df53595e3c48d24ca690d5a8
4e2029bfb2ba60a4a5bad906fb4e0bb860ee07615b854f2ae9394305f69f5f6e
3f96366c4ae531effd3f4266b231c796ba5d6d93827dab244190c49d30631ea1
null
[]
235
2.4
jims-backoffice
0.5.2
Add your description here
# JIMS UI Simple FastUI viewer of JIMS threads
text/markdown
null
Andrey Tatarinov <a@tatarinov.co>
null
null
null
null
[]
[]
null
null
>=3.12
[]
[]
[]
[ "asyncpg>=0.31.0", "fastapi>=0.115.0", "fastui>=0.7.0", "jims-core>=0.5.1", "psycopg2-binary>=2.9.10", "pydantic-settings>=2.10.1", "pydantic==2.9.2", "python-multipart>=0.0.18", "sqlalchemy>=2.0.41", "uvicorn>=0.29.0" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:06:24.376354
jims_backoffice-0.5.2.tar.gz
9,187
17/65/77de3bc9bc093ed77645c90fde8728c4f8b53e3166a065e0165b16e3fd06/jims_backoffice-0.5.2.tar.gz
source
sdist
null
false
4ea78e000bf6bd7a978ca1e76dff6868
a0b2be9b983c3c5804edb5b9fd306008f473daf4275b49f34e9a440acd3635b0
176577de3bc9bc093ed77645c90fde8728c4f8b53e3166a065e0165b16e3fd06
null
[]
243
2.1
MuiceBot
1.1.5
Muice-Chatbot 的 NoneBot2 实现
<div align=center> <img width=200 src="https://bot.snowy.moe/logo.png" alt="image"/> <h1 align="center">MuiceBot</h1> <p align="center">Muice-Chatbot 的 NoneBot2 实现</p> </div> <div align=center> <a href="#关于️"><img src="https://img.shields.io/github/stars/Moemu/MuiceBot" alt="Stars"></a> <a href="https://pypi.org/project/MuiceBot/"><img src="https://img.shields.io/pypi/v/Muicebot" alt="PyPI Version"></a> <a href="https://pypi.org/project/MuiceBot/"><img src="https://img.shields.io/pypi/dm/Muicebot" alt="PyPI Downloads" ></a> <a href="https://nonebot.dev/"><img src="https://img.shields.io/badge/nonebot-2-red" alt="nonebot2"></a> <a href="#"><img src="https://img.shields.io/badge/Code%20Style-Black-121110.svg" alt="codestyle"></a> </div> <div align=center> <a href="#"><img src="https://wakatime.com/badge/user/637d5886-8b47-4b82-9264-3b3b9d6add67/project/a4557f7b-4d26-4105-842a-7a783cbad588.svg" alt="wakatime"></a> <a href="https://www.modelscope.cn/datasets/Moemuu/Muice-Dataset"><img src="https://img.shields.io/badge/ModelScope-Dataset-644cfd?link=https://www.modelscope.cn/datasets/Moemuu/Muice-Dataset" alt="ModelScope"></a> </div> <div align=center> <a href="https://bot.snowy.moe">📃使用文档</a> <a href="https://bot.snowy.moe/guide/setup.html">✨快速开始</a> <a href="https://github.com/MuikaAI/Muicebot-Plugins-Index">🧩插件商店</a> </div> *本项目目前处于慢更新状态,目前只进行 Bug 修复/接受外部 PR 的活动,暂无进行其他功能性开发的计划* # 介绍✨ > 我们认为,AI的创造应该是为了帮助人类更好的解决问题而不是产生问题。因此,我们注重大语言模型解决实际问题的能力,如果沐雪系列项目不能帮助我们解决日常、情感类的问题,沐雪的存在就是毫无意义可言。 > *———— 《沐雪系列模型评测标准》* Muicebot 是基于 Nonebot2 框架实现的 LLM 聊天机器人,旨在解决现实问题。通过 Muicebot ,你可以在主流聊天平台(如 QQ)获得只有在网页中才能获得的聊天体验。 Muicebot 内置两个分别名为沐雪和沐妮卡的聊天人设(人设是可选的)以便优化对话体验。有关沐雪和沐妮卡的设定,还请移步 [关于沐雪](https://bot.snowy.moe/about/Muice) # 功能🪄 ✅ 内嵌多种模型加载器,如[OpenAI](https://platform.openai.com/docs/overview) 和 [Ollama](https://ollama.com/) ,可加载市面上大多数的模型服务或本地模型,支持多模态(图片识别)和工具调用。另外还附送只会计算 3.9 > 3.11 的沐雪 Roleplay 微调模型一枚~ ✅ 使用 `nonebot_plugin_alconna` 作为通用信息接口,支持市面上的大多数适配器。对部分适配器做了特殊优化 ✅ 支持基于 `nonebot_plugin_apscheduler` 的定时任务,可定时向大语言模型交互或直接发送信息 ✅ 支持基于 `nonebot_plugin_alconna` 的几条常见指令。 ✅ 基于 `nonebot-plugin-orm>=0.7.7` 提供的 ORM 层保存对话数据。那有人就要问了:Maintainer,Maintainer,能不能实现长期短期记忆、LangChain、FairSeq 这些记忆优化啊。~~以后会有的(~~ ✅ 使用 Jinja2 动态生成人设提示词 ✅ 支持调用 MCP 服务(支持 stdio、SSE 和 Streamable HTTP 传输方式) # 模型加载器适配情况 | 模型加载器 | 流式对话 | 多模态输入/输出 | 推理模型调用 | 工具调用 | 联网搜索 | | ----------- | -------- | -------------- | ------------ | -------------------- | -------------------- | | `Azure` | ✅ | 🎶🖼️/❌ | ⭕ | ✅ | ❌ | | `Dashscope` | ✅ | 🎶🖼️/❌ | ✅ | ⭕ | ✅ | | `Gemini` | ✅ | ✅/🖼️ | ⭕ | ✅ | ✅ | | `Ollama` | ✅ | 🖼️/❌ | ✅ | ✅ | ❌ | | `Openai` | ✅ | ✅/🎶 | ✅ | ✅ | ❌ | ✅:表示此加载器能很好地支持该功能并且 `MuiceBot` 已实现 ⭕:表示此加载器虽支持该功能,但使用时可能遇到问题 🚧:表示此加载器虽然支持该功能,但 `MuiceBot` 未实现或正在实现中 ❓:表示 Maintainer 暂不清楚此加载器是否支持此项功能,可能需要进一步翻阅文档和检查源码 ❌:表示此加载器不支持该功能 多模态标记:🎶表示音频;🎞️ 表示视频;🖼️ 表示图像;📄表示文件;✅ 表示完全支持 # 本项目适合谁? - 拥有编写过 Python 程序经验的开发者 - 搭建过 Nonebot 项目的 bot 爱好者 - 想要随时随地和大语言模型交互并寻找着能够同时兼容市面上绝大多数 SDK 的机器人框架的 AI 爱好者 ~~# TODO📝~~ ~~近期更新路线:[MuiceBot 更新计划](https://github.com/users/Moemu/projects/2)~~ # 使用教程💻 参考 [使用文档](https://bot.snowy.moe) # 插件商店🧩 [MuikaAI/Muicebot-Plugins-Index](https://github.com/MuikaAI/Muicebot-Plugins-Index) # 关于🎗️ 大模型输出结果将按**原样**提供,由于提示注入攻击等复杂的原因,模型有可能输出有害内容。无论模型输出结果如何,模型输出结果都无法代表开发者的观点和立场。对于此项目可能间接引发的任何后果(包括但不限于机器人账号封禁),本项目所有开发者均不承担任何责任。 本项目基于 [BSD 3](https://github.com/Moemu/nonebot-plugin-muice/blob/main/LICENSE) 许可证提供,涉及到再分发时请保留许可文件的副本。 本项目标识使用了 [nonebot/nonebot2](https://github.com/nonebot/nonebot2) 和 画师 [Nakkar](https://www.pixiv.net/users/28246124) [Pixiv作品](https://www.pixiv.net/artworks/101063891) 的资产或作品。如有侵权,请及时与我们联系 BSD 3 许可证同样适用于沐雪的系统提示词,沐雪的文字人设或人设图在 [CC BY NC 3.0](https://creativecommons.org/licenses/by-nc-sa/4.0/legalcode.zh-hans) 许可证条款下提供。 此项目中基于或参考了部分开源项目的实现,在这里一并表示感谢: - [nonebot/nonebot2](https://github.com/nonebot/nonebot2) 本项目使用的机器人框架 - [@botuniverse](https://github.com/botuniverse) 负责制定 Onebot 标准的组织 感谢各位开发者的协助,可以说没有你们就没有沐雪的今天: <a href="https://github.com/eryajf/Moemu/MuiceBot/contributors"> <img src="https://contrib.rocks/image?repo=Moemu/MuiceBot" alt="图片加载中..."/> </a> 友情链接:[LiteyukiStudio/nonebot-plugin-marshoai](https://github.com/LiteyukiStudio/nonebot-plugin-marshoai) 本项目隶属于 MuikaAI 基于 OneBot V11 的原始实现:[Moemu/Muice-Chatbot](https://github.com/Moemu/Muice-Chatbot) <a href="https://www.afdian.com/a/Moemu" target="_blank"><img src="https://pic1.afdiancdn.com/static/img/welcome/button-sponsorme.png" alt="afadian" style="height: 45px !important;width: 163px !important;"></a> <a href="https://www.buymeacoffee.com/Moemu" target="_blank"><img src="https://cdn.buymeacoffee.com/buttons/v2/default-yellow.png" alt="Buy Me A Coffee" style="height: 45px !important;width: 163px !important;" ></a> Star History: [![Star History Chart](https://api.star-history.com/svg?repos=Moemu/MuiceBot&type=Date)](https://star-history.com/#Moemu/MuiceBot&Date)
text/markdown
null
Moemu <i@snowy.moe>
null
null
null
null
[]
[]
null
null
<3.14,>=3.10.0
[]
[]
[]
[ "aiosqlite>=0.17.0", "APScheduler>=3.11.0", "fleep>=1.0.1", "jinja2>=3.1.6", "nonebot2>=2.4.1", "nonebot-adapter-onebot>=2.4.6", "nonebot_plugin_alconna>=0.57.6", "nonebot_plugin_apscheduler>=0.5.0", "nonebot_plugin_localstore>=0.7.3", "nonebot-plugin-orm>=0.7.7", "nonebot_plugin_session>=0.3.2"...
[]
[]
[]
[]
pdm/2.26.6 CPython/3.11.14 Linux/6.14.0-1017-azure
2026-02-19T12:06:06.129308
muicebot-1.1.5.tar.gz
72,752
f7/1f/f89da364e17a16f562ba26ac8a5814252d8f222384be0437637addec34f4/muicebot-1.1.5.tar.gz
source
sdist
null
false
33f20fadde533f2cd1bb0de816f0c5b4
05db7ad1b61e1f19edaefeea5b5f2d68a608773491362d05e6e412535375b386
f71ff89da364e17a16f562ba26ac8a5814252d8f222384be0437637addec34f4
null
[]
0
2.4
moltrust
0.1.0
MolTrust SDK - Trust Layer for the Agent Economy
# MolTrust SDK **Trust Layer for the Agent Economy.** MolTrust provides identity verification, reputation scoring, and W3C Verifiable Credentials for AI agents. ## Install ```bash pip install moltrust ``` ## Quickstart ```python from moltrust import MolTrust mt = MolTrust(api_key="mt_your_key") # Register an agent agent = mt.register("MyAgent") print(agent.did) # did:moltrust:a1b2c3d4... # Issue a Verifiable Credential vc = mt.issue_credential(agent.did) print(vc.is_signed) # True # Verify the credential result = mt.verify_credential(vc) print(result.valid) # True ``` ## Features - **Identity** — Register, verify, and resolve agent DIDs - **Reputation** — Rate agents and query trust scores - **Verifiable Credentials** — Issue and verify W3C VCs with Ed25519 signatures - **Async Support** — Full async client via `AsyncMolTrust` ## Integration Examples ### LangChain Tool ```python from langchain.tools import tool from moltrust import MolTrust mt = MolTrust(api_key="mt_...") @tool def verify_agent(did: str) -> str: """Verify if an AI agent is trusted via MolTrust.""" if mt.verify(did): rep = mt.get_reputation(did) return f"Verified. Trust score: {rep.score}/5 ({rep.total_ratings} ratings)" return "Agent not found." ``` ### Pre-Transaction Check ```python mt = MolTrust(api_key="mt_...") def safe_transact(counterparty_did: str): rep = mt.get_reputation(counterparty_did) if not rep.is_trusted: raise Exception(f"Not trusted (score: {rep.score})") vc = mt.issue_credential(counterparty_did, "TransactionCredential") return vc ``` ## Standards - **W3C DID:web** — Decentralized Identifiers - **W3C Verifiable Credentials** — Tamper-proof credentials - **Ed25519** — Elliptic curve signatures - **Lightning Network** — Bitcoin L2 payments ## Links - **API Docs:** https://api.moltrust.ch/docs - **DID Document:** https://api.moltrust.ch/.well-known/did.json - **Website:** https://moltrust.ch - **X:** [@moltrust](https://x.com/moltrust) ## License MIT — CryptoKRI GmbH, Zurich, Switzerland
text/markdown
null
CryptoKRI GmbH <kersten.kroehl@cryptokri.ch>
null
null
MIT
ai, agents, trust, identity, did, verifiable-credentials, reputation
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Topic :: Software Development :: Libraries", "Topic :: Security" ]
[]
null
null
>=3.9
[]
[]
[]
[ "httpx>=0.25.0" ]
[]
[]
[]
[ "Homepage, https://moltrust.ch", "Documentation, https://api.moltrust.ch/docs", "Repository, https://github.com/MoltyCel/moltrust-sdk" ]
twine/6.2.0 CPython/3.12.3
2026-02-19T12:05:56.298744
moltrust-0.1.0.tar.gz
5,787
0a/e3/0ca2235ef0a5004e0cbf2f3964fc166611368c43b278abe1116057059d8d/moltrust-0.1.0.tar.gz
source
sdist
null
false
2dab5a2b2cc79086b7910e0982f5926e
196f2e597d777c532d22728e6c0375abf6c937fbf2bcb27062e875e86fc8765e
0ae30ca2235ef0a5004e0cbf2f3964fc166611368c43b278abe1116057059d8d
null
[ "LICENSE" ]
246
2.3
dg-kit
0.1.2
Data Governance Kit provides programmatic access to data governance metadata with integrations for different Logical and Physical Modeling tools.
## Data Governance Kit (dg_kit) Data Governance Kit helps you access Data Governance information programmatically. It provides core objects that model Physical Model, Logical Model, Business Information, and related governance metadata. Integrations let you pull this data from tools like dbt, Oracle Data Modeler, and Notion, with more connectors planned in upcoming releases. This toolkit is handy for building Data Governance CI gates, strengthening Data Ops practices, and keeping governance checks close to your delivery workflows. ## Requirements - Python >= 3.10 ## Install ```bash pip install -e . ``` Optional extras: ```bash pip install -e ".[dbt]" pip install -e ".[notion]" ``` ## Quick Start ### Parse an Oracle Data Modeler project ```python from dg_kit.integrations.odm.parser import ODMParser parser = ODMParser("path/to/model.dmd") bi = parser.parse_bi() lm = parser.parse_lm() print(lm.version, len(lm.entities)) ``` ### Parse a dbt project into a physical model ```python from dg_kit.integrations.dbt.parser import DBTParser pm = DBTParser("path/to/dbt_project").parse_pm() print(pm.version, len(pm.tables)) ``` ### Validate with conventions ```python from dg_kit.base.convention import Convention, ConventionValidator from dg_kit.base.enums import ConventionRuleSeverity convention = Convention("example") @convention.rule( name="has-entities", severity=ConventionRuleSeverity.ERROR, description="Logical model must contain at least one entity", ) def has_entities(lm, pm): return set() if lm.entities else {("no entities")} issues = ConventionValidator(lm, pm, convention).validate() ``` ### Sync to Notion data catalog ```python from dg_kit.integrations.notion.api import NotionDataCatalog catalog = NotionDataCatalog( notion_token="secret", dc_table_id="data_source_id", ) rows = catalog.pull() print(len(rows)) ``` ## Development Run tests: ```bash pytest ``` Export requirements with uv: ```bash uv export --extra dbt --extra notion --group test -o requirements.txt ```
text/markdown
Chelidze Georgii
Chelidze Georgii <chelidze.georgii.d@gmail.com>
null
null
MIT
null
[]
[]
null
null
>=3.10
[]
[]
[]
[ "pyyaml>=6.0.3", "pyyaml>=6.0.3; extra == \"dbt\"", "notion-client>=2.7.0; extra == \"notion\"" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:05:10.438729
dg_kit-0.1.2.tar.gz
21,855
20/c3/3e41dec361895eb4f7fda24eb9a6736d20931edd63644e55cac0161ff89c/dg_kit-0.1.2.tar.gz
source
sdist
null
false
2e5b68e4685ab4db4b00d77463c062c5
278448097041833760debf4fe82468eb82bc21d03e5340b79be9954f5b20d73e
20c33e41dec361895eb4f7fda24eb9a6736d20931edd63644e55cac0161ff89c
null
[]
249
2.4
appwrite
15.2.0
Appwrite is an open-source self-hosted backend server that abstracts and simplifies complex and repetitive development tasks behind a very simple REST API
# Appwrite Python SDK ![License](https://img.shields.io/github/license/appwrite/sdk-for-python.svg?style=flat-square) ![Version](https://img.shields.io/badge/api%20version-1.8.1-blue.svg?style=flat-square) [![Build Status](https://img.shields.io/travis/com/appwrite/sdk-generator?style=flat-square)](https://travis-ci.com/appwrite/sdk-generator) [![Twitter Account](https://img.shields.io/twitter/follow/appwrite?color=00acee&label=twitter&style=flat-square)](https://twitter.com/appwrite) [![Discord](https://img.shields.io/discord/564160730845151244?label=discord&style=flat-square)](https://appwrite.io/discord) **This SDK is compatible with Appwrite server version 1.8.x. For older versions, please check [previous releases](https://github.com/appwrite/sdk-for-python/releases).** Appwrite is an open-source backend as a service server that abstracts and simplifies complex and repetitive development tasks behind a very simple to use REST API. Appwrite aims to help you develop your apps faster and in a more secure way. Use the Python SDK to integrate your app with the Appwrite server to easily start interacting with all of Appwrite backend APIs and tools. For full API documentation and tutorials go to [https://appwrite.io/docs](https://appwrite.io/docs) ![Appwrite](https://github.com/appwrite/appwrite/raw/main/public/images/github.png) ## Installation To install via [PyPI](https://pypi.org/): ```bash pip install appwrite ``` ## Getting Started ### Init your SDK Initialize your SDK with your Appwrite server API endpoint and project ID which can be found on your project settings page and your new API secret Key from project's API keys section. ```python from appwrite.client import Client from appwrite.services.users import Users client = Client() (client .set_endpoint('https://[HOSTNAME_OR_IP]/v1') # Your API Endpoint .set_project('5df5acd0d48c2') # Your project ID .set_key('919c2d18fb5d4...a2ae413da83346ad2') # Your secret API key .set_self_signed() # Use only on dev mode with a self-signed SSL cert ) ``` ### Make Your First Request Once your SDK object is set, create any of the Appwrite service objects and choose any request to send. Full documentation for any service method you would like to use can be found in your SDK documentation or in the [API References](https://appwrite.io/docs) section. ```python users = Users(client) result = users.create(ID.unique(), email = "email@example.com", phone = "+123456789", password = "password", name = "Walter O'Brien") ``` ### Full Example ```python from appwrite.client import Client from appwrite.services.users import Users from appwrite.id import ID client = Client() (client .set_endpoint('https://[HOSTNAME_OR_IP]/v1') # Your API Endpoint .set_project('5df5acd0d48c2') # Your project ID .set_key('919c2d18fb5d4...a2ae413da83346ad2') # Your secret API key .set_self_signed() # Use only on dev mode with a self-signed SSL cert ) users = Users(client) result = users.create(ID.unique(), email = "email@example.com", phone = "+123456789", password = "password", name = "Walter O'Brien") ``` ### Error Handling The Appwrite Python SDK raises `AppwriteException` object with `message`, `code` and `response` properties. You can handle any errors by catching `AppwriteException` and present the `message` to the user or handle it yourself based on the provided error information. Below is an example. ```python users = Users(client) try: result = users.create(ID.unique(), email = "email@example.com", phone = "+123456789", password = "password", name = "Walter O'Brien") except AppwriteException as e: print(e.message) ``` ### Learn more You can use the following resources to learn more and get help - 🚀 [Getting Started Tutorial](https://appwrite.io/docs/getting-started-for-server) - 📜 [Appwrite Docs](https://appwrite.io/docs) - 💬 [Discord Community](https://appwrite.io/discord) - 🚂 [Appwrite Python Playground](https://github.com/appwrite/playground-for-python) ## Contribution This library is auto-generated by Appwrite custom [SDK Generator](https://github.com/appwrite/sdk-generator). To learn more about how you can help us improve this SDK, please check the [contribution guide](https://github.com/appwrite/sdk-generator/blob/master/CONTRIBUTING.md) before sending a pull-request. ## License Please see the [BSD-3-Clause license](https://raw.githubusercontent.com/appwrite/appwrite/master/LICENSE) file for more information.
text/markdown
Appwrite Team
team@appwrite.io
Appwrite Team
team@appwrite.io
BSD-3-Clause
null
[ "Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Environment :: Web Environment", "Topic :: Software Development", "License :: OSI Approved :: BSD License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python...
[]
https://appwrite.io/support
https://github.com/appwrite/sdk-for-python/archive/15.2.0.tar.gz
null
[]
[]
[]
[ "requests" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.9.25
2026-02-19T12:04:39.156179
appwrite-15.2.0.tar.gz
78,695
86/d3/5e3ccb9dc8d36353836bfdba17864768718e3cdff62c0748fdf76b47b812/appwrite-15.2.0.tar.gz
source
sdist
null
false
586c5d69487316c2a9a414ff58d5310b
05dffcadceac4a909e5c30fdf294a01ecaa810b1bdca86e1fdbe39f775adf309
86d35e3ccb9dc8d36353836bfdba17864768718e3cdff62c0748fdf76b47b812
null
[ "LICENSE" ]
3,987
2.4
pyremotedata
0.1.7
A package for low- and high-level high-bandwidth asynchronous data transfer
# `pyRemoteData` `pyRemoteData` is a module developed for scientific computation using the remote storage platform [ERDA](https://erda.au.dk/) (Electronic Research Data Archive) provided by Aarhus University IT, as part of my PhD at the Department of Ecoscience at Aarhus University. It can be used with **any** storage facility that supports SFTP and LFTP, but is only tested on a minimal SFTP server found at [atmoz/sftp](https://hub.docker.com/r/atmoz/sftp) and on the live AU ERDA service which runs on MiG (Minimum intrusion Grid - [SourceForge](https://sourceforge.net/projects/migrid/)/[GitHub](https://github.com/ucphhpc/migrid-sync)) developed by [SCIENCE HPC Centre at Copenhagen University](https://science.ku.dk/english/research/research-e-infrastructure/science-hpc-centre/). ## Capabilities In order to facility high-throughput computation in a cross-platform setting, `pyRemoteData` handles data transfer with multithreading and asynchronous data streaming using thread-safe buffers. ## Use-cases If your storage facility supports SFTP and LFTP, and you need high-bandwidth data streaming for analysis, data migration or other purposes such as model-training, then this module may be of use to you. Experience with SFTP or LFTP is not necessary, but you must be able to setup the required SSH configurations. See **Automated** for details on how to avoid having to set up SSH configuration. ## Setup A more user-friendly setup process, which facilitates both automated as well as interactive setup is currently in development. (**TODO**: Finish and describe the setup process) ### Installation The package is available on PyPI, and can be installed using pip: ```bash pip install pyremotedata ``` ### Interactive Simply follow the popup instructions that appear once you load the package for the first time. ### Automated The automatic configuration setup relies on setting the correct environment variables **BEFORE LOADING THE PACKAGE**: * `PYREMOTEDATA_REMOTE_USERNAME` : Should be set to your username on your remote service. * `PYREMOTEDATA_REMOTE_URI` : Should be set to the URI of the endpoint for your remote service (e.g. for ERDA it is "io.erda.au.dk"). * `PYREMOTEDATA_REMOTE_DIRECTORY` : If you would like to set a default working directory, that is not the root of your remote storage, then set this to that (e.g. "/MY_PROJECT/DATASETS") otherwise simply set this to "/". * `PYREMOTEDATA_AUTO` : Should be **set to "yes"** to disable interactive mode. If this is not set, or set to anything other than "yes" (not case-sensitive), while any of the prior environment variables are unset an error will be thrown. The recommended way to avoid any SSH or environment variables setup is to use: ```py from pyremotedata.implicit_mount import IOHandler with IOHandler(lftp_settings = {'sftp:connect-program' : 'ssh -a -x -i <keyfile>'}, user = <USER>, remote = <REMOTE>) as io: ... ``` Here `keyfile` is probably something like `~/.ssh/id_rsa`. ### Example If you want to test against a mock server simply follow the instructions in tests/README. If you have a remote storage facility that supports SFTP and LFTP, then you can use the following example to test the functionality of the module: ```python # Set the environment variables (only necessary in a non-interactive setting) # If you are simply running this as a Python script, # you can omit these lines and you will be prompted to set them interactively import os os.environ["PYREMOTEDATA_REMOTE_USERNAME"] = "username" os.environ["PYREMOTEDATA_REMOTE_URI"] = "storage.example.com" os.environ["PYREMOTEDATA_REMOTE_DIRECTORY"] = "/MY_PROJECT/DATASETS" os.environ["PYREMOTEDATA_AUTO"] = "yes" from pyremotedata.implicit_mount import IOHandler handler = IOHandler() with handler as io: print(io.ls()) local_file = io.download("/remote/file/or/directory") # The configuration is persistent, but can be removed using the following: from pyremotedata.config import remove_config remove_config() ``` ## Issues This module is certainly not maximally efficient, and you may run into network- or OS-specific issues. Any and all feedback and contributions is highly appreciated.
text/markdown
null
Asger Svenning <asgersvenning@gmail.com>
null
null
null
null
[ "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3" ]
[]
null
null
>=3.10
[]
[]
[]
[ "pyyaml>=6.0.1", "tqdm>=4.67.1", "colorlog>=6.9.0; extra == \"full\"", "torch>=2.0.1; extra == \"full\"", "torchvision>=0.15.2; extra == \"full\"", "wrapt-timeout-decorator>=1.5.1; extra == \"full\"" ]
[]
[]
[]
[ "Homepage, https://github.com/asgersvenning/pyremotedata", "Bug Tracker, https://github.com/asgersvenning/pyremotedata/issues" ]
twine/6.2.0 CPython/3.10.19
2026-02-19T12:04:12.527352
pyremotedata-0.1.7.tar.gz
51,253
08/08/cc946b8153fd46343ee310d78ae8b9d56d7de81cb66b5b71855ef3c65c47/pyremotedata-0.1.7.tar.gz
source
sdist
null
false
32347e943d3341148f2d37c01439ab64
8bdb1c7d19a8e512d5454d24df321d74df8618f3c7fd390f9c3c6de393127504
0808cc946b8153fd46343ee310d78ae8b9d56d7de81cb66b5b71855ef3c65c47
null
[ "LICENSE" ]
240
2.2
xtgeo
4.17.1
XTGeo is a Python library for 3D grids, surfaces, wells, etc
![XTGeo](https://github.com/equinor/xtgeo/blob/main/docs/images/xtgeo-logo-wide.png) ![builds](https://github.com/equinor/xtgeo/workflows/builds/badge.svg) ![linting](https://github.com/equinor/xtgeo/workflows/linting/badge.svg) [![codecov](https://codecov.io/gh/equinor/xtgeo/branch/main/graph/badge.svg)](https://codecov.io/gh/equinor/xtgeo) [![Ruff](https://img.shields.io/endpoint?url=https://raw.githubusercontent.com/astral-sh/ruff/main/assets/badge/v2.json)](https://github.com/astral-sh/ruff) [![PyPI version](https://badge.fury.io/py/xtgeo.svg)](https://badge.fury.io/py/xtgeo) [![Documentation Status](https://readthedocs.org/projects/xtgeo/badge/?version=latest)](https://xtgeo.readthedocs.io/en/latest/?badge=latest) ![PyPI - Python Version](https://img.shields.io/pypi/pyversions/xtgeo.svg) ![PyPI - License](https://img.shields.io/pypi/l/xtgeo.svg) ## Introduction XTGeo is a LGPL licensed Python library with C backend to support manipulation of (oil industry) subsurface reservoir modelling. Typical users are geoscientist and reservoir engineers working with reservoir modelling, in relation with RMS. XTGeo is developed in Equinor. Detailed documentation for [XTGeo at Read _the_ Docs](https://xtgeo.readthedocs.io) ## Feature summary - Python 3.10+ support - Focus on high speed, using numpy and pandas with C backend - Regular surfaces, i.e. 2D maps with regular sampling and rotation - 3D grids (corner-point), supporting several formats such as RMS and Eclipse - Support of seismic cubes, using [segyio](https://github.com/equinor/segyio) as backend for SEGY format - Support of well data, line and polygons (still somewhat immature) - Operations between the data types listed above; e.g. slice a surface with a seismic cube - Optional integration with ROXAR API python for several data types (see note later) - Linux is main development platform, but Windows and MacOS (64 bit) are supported and PYPI wheels for all three platforms are provided. ## Installation For Linux, Windows and MacOS 64bit, PYPI installation is enabled: ``` pip install xtgeo ``` For detailed installation instructions (implies C compiling), see the documentation. ## Getting started ```python import xtgeo # create an instance of a surface, read from file mysurf = xtgeo.surface_from_file("myfile.gri") # Irap binary as default print(f"Mean is {mysurf.values.mean()}") # change date so all values less than 2000 becomes 2000 # The values attribute gives the Numpy array mysurface.values[mysurface.values < 2000] = 2000 # export the modified surface: mysurface.to_file("newfile.gri") ``` ## Note on RMS Roxar API integration The following applies to the part of the XTGeo API that is connected to Roxar API (RMS): > RMS is neither an open source software nor a free software and > any use of it needs a software license agreement in place.
text/markdown
null
Equinor <fg_fmu-atlas@equinor.com>
null
null
LGPL-3.0
grids, surfaces, wells, cubes
[ "Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: Science/Research", "License :: OSI Approved :: GNU Lesser General Public License v3 (LGPLv3)", "Operating System :: POSIX :: Linux", "Operating System :: Microsoft :: Windows", "Operating System :: Ma...
[]
null
null
>=3.10
[]
[]
[]
[ "deprecation", "gstools", "h5py>=3", "hdf5plugin>=2.3", "joblib", "matplotlib>=3.3", "numpy", "pandas<3.0.0,>=1.1", "pyarrow", "resfo>=5", "roffio>=0.0.2", "scipy>=1.5", "segyio>1.8.0", "shapely>=1.6.2", "tables", "typing_extensions", "xtgeoviz", "clang-format; extra == \"dev\"", ...
[]
[]
[]
[ "Homepage, https://github.com/equinor/xtgeo", "Repository, https://github.com/equinor/xtgeo", "Issues, https://github.com/equinor/xtgeo/issues", "Documentation, https://xtgeo.readthedocs.io" ]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:04:11.577786
xtgeo-4.17.1-cp313-cp313-win_amd64.whl
1,287,565
93/53/aa2ab6ccc24fa7adabcc8b6e41550dde00cf078e88796ac5dea7175ca032/xtgeo-4.17.1-cp313-cp313-win_amd64.whl
cp313
bdist_wheel
null
false
33030dad09517a42da6cfcbe76bf3a23
656edb0fb53652a7a58fd7378ae45a5ab45b5a9e0da5918483e5fa08aefe3fd7
9353aa2ab6ccc24fa7adabcc8b6e41550dde00cf078e88796ac5dea7175ca032
null
[]
1,571
2.4
mdify-cli
3.6.5
Convert PDFs and document images into structured Markdown for LLM workflows
# mdify ![mdify banner](https://raw.githubusercontent.com/tiroq/mdify/main/assets/mdify.png) [![PyPI](https://img.shields.io/pypi/v/mdify-cli?logo=python&style=flat-square)](https://pypi.org/project/mdify-cli/) [![Container](https://img.shields.io/badge/container-ghcr.io-blue?logo=docker&style=flat-square)](https://github.com/tiroq/mdify/pkgs/container/mdify-runtime) [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg?style=flat-square)](https://opensource.org/licenses/MIT) A lightweight CLI for converting documents to Markdown. The CLI is fast to install via pipx, while the heavy ML conversion runs inside a container. ## Requirements - **Python 3.8+** - **Docker**, **Podman**, or native macOS container tools (for document conversion) - On macOS: Supports Apple Container (macOS 26+), OrbStack, Colima, Podman, or Docker Desktop - On Linux: Docker or Podman - Auto-detects available tools ## Installation ### macOS (recommended) ```bash brew install pipx pipx ensurepath pipx install mdify-cli ``` Restart your terminal after installation. For containerized document conversion, install one of these (or use Docker Desktop): - **Apple Container** (macOS 26+): Download from https://github.com/apple/container/releases - **OrbStack** (recommended): `brew install orbstack` - **Colima**: `brew install colima && colima start` - **Podman**: `brew install podman && podman machine init && podman machine start` - **Docker Desktop**: Available at https://www.docker.com/products/docker-desktop ### Linux ```bash python3 -m pip install --user pipx pipx ensurepath pipx install mdify-cli ``` ### Install via pip ```bash pip install mdify-cli ``` ### Development install ```bash git clone https://github.com/tiroq/mdify.git cd mdify pip install -e . ``` ## Usage ### Basic conversion Convert a single file: ```bash mdify document.pdf ``` The first run will automatically pull the container image (~2GB) if not present. ### Convert multiple files Convert all PDFs in a directory: ```bash mdify /path/to/documents -g "*.pdf" ``` Recursively convert files: ```bash mdify /path/to/documents -r -g "*.pdf" ``` ### GPU Acceleration For faster processing with NVIDIA GPU: ```bash mdify --gpu documents/*.pdf ``` Requires NVIDIA GPU with CUDA support and nvidia-container-toolkit. ### 🚀 Remote Server Execution (SSH) **NEW:** Convert documents on remote servers via SSH to offload resource-intensive processing: ```bash # Basic remote conversion mdify document.pdf --remote-host server.example.com # Use SSH config alias mdify document.pdf --remote-host production # With custom configuration mdify docs/*.pdf --remote-host 192.168.1.100 \ --remote-user admin \ --remote-key ~/.ssh/id_rsa # Validate remote server before processing mdify document.pdf --remote-host server --remote-validate-only ``` **How it works:** 1. Connects to remote server via SSH 2. Validates remote resources (disk space, memory, Docker/Podman) 3. Uploads files via SFTP 4. Starts remote container automatically 5. Converts documents on remote server 6. Downloads results via SFTP 7. Cleans up remote files and stops container **Requirements:** - SSH key authentication (password auth not supported for security) - Docker or Podman installed on remote server - Minimum 5GB disk space and 2GB RAM on remote **SSH Configuration:** Create `~/.mdify/remote.conf` for reusable settings: ```yaml host: production.example.com port: 22 username: deploy key_file: ~/.ssh/deploy_key work_dir: /tmp/mdify-remote container_runtime: docker timeout: 30 ``` Or use existing `~/.ssh/config`: ``` Host production HostName 192.168.1.100 User deploy Port 2222 IdentityFile ~/.ssh/deploy_key ``` Then simply: `mdify doc.pdf --remote-host production` **Configuration Precedence** (highest to lowest): 1. CLI arguments (`--remote-*`) 2. `~/.mdify/remote.conf` 3. `~/.ssh/config` 4. Built-in defaults See the [SSH Remote Server Guide](#ssh-remote-server-options) below for all options. ### ⚠️ PII Masking (Deprecated) The `--mask` flag is deprecated and will be ignored in this version. PII masking functionality was available in older versions using a custom runtime but is not supported with the current docling-serve backend. If PII masking is critical for your use case, please use mdify v1.5.x or earlier versions. ## Performance mdify now uses docling-serve for significantly faster batch processing: - **Single model load**: Models are loaded once per session, not per file - **~10-20x speedup** for multiple file conversions compared to previous versions - **GPU acceleration**: Use `--gpu` for additional 2-6x speedup (requires NVIDIA GPU) ### First Run Behavior The first conversion takes longer (~30-60s) as the container loads ML models into memory. Subsequent files in the same batch process quickly, typically in 1-3 seconds per file. ## Options | Option | Description | |--------|-------------| | `input` | Input file or directory to convert (required) | | `-o, --out-dir DIR` | Output directory for converted files (default: output) | | `-g, --glob PATTERN` | Glob pattern for filtering files (default: *) | | `-r, --recursive` | Recursively scan directories | | `--flat` | Disable directory structure preservation | | `--overwrite` | Overwrite existing output files | | `-q, --quiet` | Suppress progress messages | | `-m, --mask` | ⚠️ **Deprecated**: PII masking not supported in current version | | `--gpu` | Use GPU-accelerated container (requires NVIDIA GPU and nvidia-container-toolkit) | | `--port PORT` | Container port (default: 5001) | | `--runtime RUNTIME` | Container runtime: docker, podman, orbstack, colima, or container (auto-detected) | | `--image IMAGE` | Custom container image (default: ghcr.io/docling-project/docling-serve-cpu:main) | | `--pull POLICY` | Image pull policy: always, missing, never (default: missing) | | `--check-update` | Check for available updates and exit | | `--version` | Show version and exit | ### SSH Remote Server Options | Option | Description | | ------ | ----------- | | `--remote-host HOST` | SSH hostname or IP (required for remote mode) | | `--remote-port PORT` | SSH port (default: 22) | | `--remote-user USER` | SSH username (uses ~/.ssh/config or current user) | | `--remote-key PATH` | SSH private key file path | | `--remote-key-passphrase PASS` | SSH key passphrase | | `--remote-timeout SEC` | SSH connection timeout in seconds (default: 30) | | `--remote-work-dir DIR` | Remote working directory (default: /tmp/mdify-remote) | | `--remote-runtime RT` | Remote container runtime: docker or podman (auto-detected) | | `--remote-config PATH` | Path to mdify remote config file (default: ~/.mdify/remote.conf) | | `--remote-skip-ssh-config` | Don't load settings from ~/.ssh/config | | `--remote-skip-validation` | Skip remote resource validation (not recommended) | | `--remote-validate-only` | Validate remote server and exit (dry run) | | `--remote-debug` | Enable detailed SSH debug logging | ### Container Runtime Selection mdify automatically detects and uses the best available container runtime. The detection order differs by platform: **macOS (recommended):** 1. Apple Container (native, macOS 26+ required) 2. OrbStack (lightweight, fast) 3. Colima (open-source alternative) 4. Podman (via Podman machine) 5. Docker Desktop (full Docker) **Linux:** 1. Docker 2. Podman **Override runtime:** Use the `MDIFY_CONTAINER_RUNTIME` environment variable to force a specific runtime: ```bash export MDIFY_CONTAINER_RUNTIME=orbstack mdify document.pdf ``` Or inline: ```bash MDIFY_CONTAINER_RUNTIME=colima mdify document.pdf ``` **Supported values:** `docker`, `podman`, `orbstack`, `colima`, `container` If the selected runtime is installed but not running, mdify will display a helpful warning: ``` Warning: Found container runtime(s) but daemon is not running: - orbstack (/opt/homebrew/bin/orbstack) Please start one of these tools before running mdify. macOS tip: Start OrbStack, Colima, or Podman Desktop application ``` With `--flat`, all output files are placed directly in the output directory. Directory paths are incorporated into filenames to prevent collisions: - `docs/subdir1/file.pdf` → `output/subdir1_file.md` - `docs/subdir2/file.pdf` → `output/subdir2_file.md` ## Examples Convert all PDFs recursively, preserving structure: ```bash mdify documents/ -r -g "*.pdf" -o markdown_output ``` Convert with Podman instead of Docker: ```bash mdify document.pdf --runtime podman ``` Use a custom/local container image: ```bash mdify document.pdf --image my-custom-image:latest ``` Force pull latest container image: ```bash mdify document.pdf --pull ``` ## Architecture ``` ┌──────────────────┐ ┌─────────────────────────────────┐ │ mdify CLI │ │ Container (Docker/Podman) │ │ (lightweight) │────▶│ ┌───────────────────────────┐ │ │ │ │ │ Docling + ML Models │ │ │ - File handling │◀────│ │ - PDF parsing │ │ │ - Container │ │ │ - OCR (Tesseract) │ │ │ orchestration │ │ │ - Document conversion │ │ └──────────────────┘ │ └───────────────────────────┘ │ └─────────────────────────────────┘ ``` The CLI: - Installs in seconds via pipx (no ML dependencies) - Automatically detects Docker or Podman - Pulls the runtime container on first use - Mounts files and runs conversions in the container ## Container Images mdify uses official docling-serve containers: **CPU Version** (default): ``` ghcr.io/docling-project/docling-serve-cpu:main ``` **GPU Version** (use with `--gpu` flag): ``` ghcr.io/docling-project/docling-serve-cu126:main ``` These are official images from the [docling-serve project](https://github.com/DS4SD/docling-serve). ## Updates mdify checks for updates daily. When a new version is available: ``` ================================================== A new version of mdify is available! Current version: 0.3.0 Latest version: 0.4.0 ================================================== Run upgrade now? [y/N] ``` ### Disable update checks ```bash export MDIFY_NO_UPDATE_CHECK=1 ``` ## Uninstall ```bash pipx uninstall mdify-cli ``` Or if installed via pip: ```bash pip uninstall mdify-cli ``` ## Troubleshooting ### SSH Remote Server Issues **Connection Refused** ``` Error: SSH connection failed: Connection refused (host:22) ``` - Verify SSH server is running on remote: `ssh user@host` - Check firewall allows port 22 (or custom SSH port) - Verify hostname/IP is correct **Authentication Failed** ``` Error: SSH authentication failed ``` - Use SSH key authentication (password auth not supported) - Verify key file exists: `ls -l ~/.ssh/id_rsa` - Check key permissions: `chmod 600 ~/.ssh/id_rsa` - Test SSH manually: `ssh -i ~/.ssh/id_rsa user@host` - Add key to ssh-agent: `ssh-add ~/.ssh/id_rsa` **Remote Container Runtime Not Found** ``` Error: Container runtime not available: docker/podman ``` - Install Docker on remote: `sudo apt install docker.io` (Ubuntu/Debian) - Or install Podman: `sudo dnf install podman` (Fedora/RHEL) - Add user to docker group: `sudo usermod -aG docker $USER` - Verify remote Docker running: `ssh user@host docker ps` **Insufficient Remote Resources** ``` Warning: Less than 5GB available on remote ``` - Free up disk space on remote server - Use `--remote-work-dir` to specify different partition - Use `--remote-skip-validation` to bypass check (not recommended) **File Transfer Timeout** ``` Error: File transfer timeout ``` - Increase timeout: `--remote-timeout 120` - Check network bandwidth and stability - Try smaller files first to verify connection **Container Health Check Fails** ``` Error: Container failed to become healthy within 60 seconds ``` - Check remote Docker logs: `ssh user@host docker logs mdify-remote-<id>` - Verify port 5001 not in use: `ssh user@host netstat -tuln | grep 5001` - Try different port: `--port 5002` **SSH Config Not Loaded** If using SSH config alias but getting connection errors: ```bash # Verify SSH config is valid cat ~/.ssh/config # Test SSH config works ssh your-alias # Use explicit connection if needed mdify doc.pdf --remote-host 192.168.1.100 --remote-user admin ``` **Permission Denied on Remote** ``` Error: Work directory not writable: /tmp/mdify-remote ``` - SSH to remote and check permissions: `ssh user@host ls -ld /tmp` - Use directory in your home: `--remote-work-dir ~/mdify-temp` - Fix permissions: `ssh user@host chmod 777 /tmp/mdify-remote` **Debug Mode** Enable detailed logging for troubleshooting: ```bash # Debug SSH operations mdify doc.pdf --remote-host server --remote-debug # Debug local operations MDIFY_DEBUG=1 mdify doc.pdf ``` ## Development ### Task automation This project uses [Task](https://taskfile.dev) for automation: ```bash # Show available tasks task # Build package task build # Build container locally task container-build # Release workflow task release-patch ``` ### Building for PyPI See [PUBLISHING.md](PUBLISHING.md) for complete publishing instructions. ## License MIT
text/markdown
tiroq
null
null
null
null
markdown, conversion, pdf, docling, cli, document, docker
[ "Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Developers", "Intended Audience :: End Users/Desktop", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "P...
[]
null
null
>=3.10
[]
[]
[]
[ "requests", "asyncssh>=2.10.0", "pyyaml>=6.0", "pytest>=7.0; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://github.com/tiroq/mdify", "Repository, https://github.com/tiroq/mdify", "Issues, https://github.com/tiroq/mdify/issues" ]
twine/6.2.0 CPython/3.14.2
2026-02-19T12:04:07.031960
mdify_cli-3.6.5.tar.gz
1,866,515
5e/3c/a88dfd1c77dda77bd7e0b05cde6024a906daf503d9b1ad996f5ad31e6531/mdify_cli-3.6.5.tar.gz
source
sdist
null
false
220026eb9db290545927ed55f9551d6d
611b01fdbf63384306d75afebf6565f3b15bc53e0dd5258e58a7c8ab95452779
5e3ca88dfd1c77dda77bd7e0b05cde6024a906daf503d9b1ad996f5ad31e6531
MIT
[ "LICENSE" ]
238
2.4
rtftree
1.0.2
Terraform Project Tree Generator CLI Tool
# 🚀 rtftree --- Terraform Project Tree Generator CLI > A professional-grade CLI tool to generate clean, structured, and > shareable Terraform project trees --- with smart exclude support, > colored output, and Markdown export. ------------------------------------------------------------------------ ![Python](https://img.shields.io/badge/python-3.8+-blue.svg) ![License](https://img.shields.io/badge/license-MIT-green.svg) ![CLI](https://img.shields.io/badge/type-CLI-orange.svg) ![Terraform](https://img.shields.io/badge/terraform-supported-purple.svg) ------------------------------------------------------------------------ ## 📌 Why rtftree? Terraform projects often include: - `.terraform/` - `terraform.tfstate` - `.terraform.lock.hcl` - `.git/` - Provider binaries - Deeply nested modules Sharing structure manually becomes messy and unreadable. 🔥 **rtftree solves this problem** by generating a clean, structured tree view of your infrastructure project --- ready for documentation, sharing, and audits. ------------------------------------------------------------------------ # ✨ Features ✅ Beautiful tree-style output\ ✅ 🎨 Colored CLI output\ ✅ Optional file content preview\ ✅ Smart exclude support (like `.gitignore`)\ ✅ Wildcard pattern support (`*.exe`, `.terraform*`)\ ✅ Exclude via file (`--exclude-file`)\ ✅ Markdown export mode (`--markdown`)\ ✅ Output to file (`-o`)\ ✅ Lightweight & Fast\ ✅ Installable as a CLI tool ------------------------------------------------------------------------ # 📦 Installation ## 🔹 Local Install (Development Mode) ``` bash python -m pip install -e . ``` ## 🔹 Standard Install ``` bash python -m pip install . ``` ## 🔹 After PyPI Publish (Global Install) ``` bash pip install rtftree ``` ------------------------------------------------------------------------ # 🚀 Usage ## Basic Usage ``` bash rtftree <project-folder> ``` Example: ``` bash rtftree . ``` ------------------------------------------------------------------------ ## 📁 Structure Only (No File Content) ``` bash rtftree . --no-content ``` ------------------------------------------------------------------------ ## 🚫 Exclude Files & Folders ### Direct Patterns ``` bash rtftree . --exclude .terraform .git terraform.tfstate *.exe ``` ### Using Exclude File ``` bash rtftree . --exclude-file exclude.txt ``` Example `exclude.txt`: .terraform .git terraform.tfstate *.exe .terraform.lock.hcl Supported: - Exact file names - Folder names - Wildcards ------------------------------------------------------------------------ ## 💾 Save Output to File ``` bash rtftree . -o infra_tree.txt ``` ------------------------------------------------------------------------ ## 📝 Markdown Export Mode Generate Markdown-ready structure: ``` bash rtftree . --markdown -o structure.md ``` Perfect for: - GitHub documentation - Wiki pages - Confluence - Client documentation ------------------------------------------------------------------------ # 🖥 Example Output 📁 Terraform Project: infra ├── 📁 modules │ ├── 📄 main.tf │ │ resource "azurerm_resource_group" "rg" { │ │ name = "example" │ │ location = "East US" │ │ } │ └── 📄 variables.tf │ variable "location" { │ type = string │ } └── 📄 provider.tf ------------------------------------------------------------------------ # ⚙️ CLI Options Option Description ------------------ ------------------------------------ `--no-content` Show only folder/file structure `--exclude` Space-separated patterns to ignore `--exclude-file` Load exclude patterns from file `--markdown` Export output in Markdown format `-o`, `--output` Write output to file ------------------------------------------------------------------------ # 🏗 Project Structure rtftree/ │ ├── rtftree/ │ ├── __init__.py │ └── cli.py │ ├── pyproject.toml └── README.md ------------------------------------------------------------------------ # 🔥 DevOps Use Cases - Share Terraform structure in tickets - Infrastructure documentation - CI/CD pipeline documentation - Client infrastructure overview - Audit reporting - Pre-deployment reviews ------------------------------------------------------------------------ # 🌍 PyPI Publishing Once published to PyPI, anyone can install globally: ``` bash pip install rtftree ``` This makes rtftree a globally accessible DevOps utility tool. ------------------------------------------------------------------------ # 🧠 Roadmap - `.treeignore` auto-detection - `--max-depth` option - Terraform-only mode (`*.tf`) - JSON export - GitHub Action integration - Auto documentation mode ------------------------------------------------------------------------ # 👨‍💻 Author **Ritesh Sharma**\ DevOps \| Azure \| Terraform \| Kubernetes ------------------------------------------------------------------------ # 📄 License MIT License ------------------------------------------------------------------------ # ⭐ Support If you find this tool useful: ⭐ Star the repository\ 🚀 Share with DevOps community\ 🛠 Contribute improvements ------------------------------------------------------------------------ > Built with ❤️ for DevOps Engineers
text/markdown
Ritesh Sharma
null
null
null
MIT
null
[]
[]
null
null
>=3.8
[]
[]
[]
[ "rich>=13.0.0" ]
[]
[]
[]
[ "Homepage, https://github.com/riteshatri/rtftree" ]
twine/6.2.0 CPython/3.11.9
2026-02-19T12:04:00.363186
rtftree-1.0.2.tar.gz
5,487
2b/14/7e909c2b65a7742a68161dfff50bc3e72dfaae8bf7ddbaffb320b0a329b9/rtftree-1.0.2.tar.gz
source
sdist
null
false
0fe7af4f89414f1540514c3253974a77
420c8e746d0b513d7b6d841032b1f9e87a6ef908e69b85775915366adf5b00aa
2b147e909c2b65a7742a68161dfff50bc3e72dfaae8bf7ddbaffb320b0a329b9
null
[ "LICENSE" ]
239
2.4
parcellate
0.2.0
A simple tool for extracting regional statistics from scalar neuroimaging maps.
# parcellate [![Build status](https://img.shields.io/github/actions/workflow/status/GalKepler/parcellate/main.yml?branch=main)](https://github.com/GalKepler/parcellate/actions/workflows/main.yml?query=branch%3Amain) [![codecov](https://codecov.io/gh/GalKepler/parcellate/branch/main/graph/badge.svg)](https://codecov.io/gh/GalKepler/parcellate) [![Documentation](https://readthedocs.org/projects/neuroparcellate/badge/?version=latest)](https://neuroparcellate.readthedocs.io/en/latest/?version=latest) [![License](https://img.shields.io/github/license/GalKepler/parcellate)](https://img.shields.io/github/license/GalKepler/parcellate) > Extract regional statistics from scalar neuroimaging maps using atlas-based parcellation. ## What It Does **parcellate** is a Python tool that extracts regional statistics from volumetric brain images using atlas-based parcellation. Given a scalar map (e.g., gray matter density, fractional anisotropy) and a labeled atlas, it computes summary statistics for each brain region. **Key Features:** - 🧠 **Multiple Pipeline Support**: Integrates with CAT12 (VBM) and QSIRecon (diffusion MRI) outputs - 📊 **Rich Statistics**: 13+ built-in metrics including mean, median, volume, robust statistics, and higher-order moments - ⚡ **Performance**: Parallel processing, smart caching, and optimized resampling - 📁 **BIDS-Compatible**: Outputs follow BIDS-derivative conventions - 🔧 **Flexible**: Python API and CLI interfaces, custom atlases and statistics **Supported Input Formats:** - **CAT12**: Gray matter (GM), white matter (WM), CSF volumes, cortical thickness maps - **QSIRecon**: Diffusion scalar maps (FA, MD, AD, RD, etc.) **Output Format:** - TSV files with regional statistics (one row per brain region) - BIDS-derivative compatible directory structure ## Installation ```bash pip install parcellate ``` For CAT12 CSV batch processing with environment variable support: ```bash pip install parcellate[dotenv] ``` ## Quick Start ### CAT12 Pipeline Process CAT12 VBM outputs with a TOML configuration file: ```bash parcellate cat12 config.toml ``` **Example config.toml:** ```toml input_root = "/data/cat12_derivatives" output_dir = "/data/parcellations" subjects = ["sub-01", "sub-02"] # Optional: process specific subjects force = false # Skip existing outputs log_level = "INFO" n_jobs = 4 # Parallel jobs within subject n_procs = 2 # Parallel processes across subjects [[atlases]] name = "Schaefer400" path = "/atlases/Schaefer2018_400Parcels_7Networks_order_FSLMNI152_1mm.nii.gz" lut = "/atlases/Schaefer2018_400Parcels_7Networks_order.tsv" space = "MNI152NLin2009cAsym" ``` **Example output:** ``` parcellations/ └── sub-01/ └── anat/ ├── sub-01_space-MNI152NLin2009cAsym_atlas-Schaefer400_desc-gm_stats.tsv ├── sub-01_space-MNI152NLin2009cAsym_atlas-Schaefer400_desc-wm_stats.tsv └── sub-01_space-MNI152NLin2009cAsym_atlas-Schaefer400_desc-ct_stats.tsv ``` ### QSIRecon Pipeline Process QSIRecon diffusion outputs: ```bash parcellate qsirecon --config config.toml --input-root /data/qsirecon ``` **Example config.toml:** ```toml input_root = "/data/qsirecon_derivatives" output_dir = "/data/parcellations" sessions = ["ses-01"] n_jobs = 4 [[atlases]] name = "JHU" path = "/atlases/JHU-ICBM-labels-1mm.nii.gz" lut = "/atlases/JHU-ICBM-labels-1mm.tsv" space = "MNI152NLin6Asym" ``` ### CAT12 CSV Mode (Batch Processing) Process multiple subjects from a CSV file: ```bash parcellate-cat12 subjects.csv --root /data/cat12 --atlas-path /atlases/schaefer400.nii.gz ``` **Example subjects.csv:** ```csv subject_id,session_id sub-01,ses-baseline sub-02,ses-baseline ``` Environment variables can be used for configuration: ```bash export CAT12_ROOT=/data/cat12_derivatives export CAT12_OUTPUT_DIR=/data/parcellations export CAT12_ATLAS_PATHS=/atlases/atlas1.nii.gz,/atlases/atlas2.nii.gz export CAT12_ATLAS_NAMES=Schaefer400,AAL3 parcellate-cat12 subjects.csv ``` ## Configuration Reference ### TOML Configuration | Field | Type | Description | Default | |-------|------|-------------|---------| | `input_root` | string | Path to preprocessing derivatives | Required | | `output_dir` | string | Output directory for parcellations | `{input_root}/parcellations` | | `subjects` | list | Subject IDs to process | All discovered | | `sessions` | list | Session IDs to process | All discovered | | `mask` | string | Brain mask path or builtin (`gm`, `wm`, `brain`) | None | | `force` | boolean | Overwrite existing outputs | `false` | | `log_level` | string | Logging verbosity (`DEBUG`, `INFO`, `WARNING`) | `INFO` | | `n_jobs` | integer | Parallel jobs within subject | `1` | | `n_procs` | integer | Parallel processes across subjects | `1` | ### Atlas Definition ```toml [[atlases]] name = "MyAtlas" # Atlas identifier path = "/path/to/atlas.nii.gz" # NIfTI file with integer labels lut = "/path/to/atlas.tsv" # Optional: TSV with columns 'index' and 'label' space = "MNI152NLin2009cAsym" # Template space ``` The LUT (lookup table) TSV should have: - `index`: Integer region IDs matching atlas labels - `label`: Region names ### Environment Variables (CAT12 CSV Mode) | Variable | Description | |----------|-------------| | `CAT12_ROOT` | Input root directory (required) | | `CAT12_OUTPUT_DIR` | Output directory | | `CAT12_ATLAS_PATHS` | Comma-separated atlas paths | | `CAT12_ATLAS_NAMES` | Comma-separated atlas names | | `CAT12_ATLAS_SPACE` | Space for all atlases | | `CAT12_MASK` | Mask path or builtin name | | `CAT12_LOG_LEVEL` | Logging level | ## Output Format Each parcellation produces a TSV file with one row per brain region: | Column | Description | |--------|-------------| | `index` | Region ID from atlas | | `label` | Region name (if LUT provided) | | `mean` | Mean intensity | | `std` | Standard deviation | | `median` | Median intensity | | `mad_median` | Median absolute deviation | | `min` | Minimum intensity | | `max` | Maximum intensity | | `range` | Max - Min | | `volume` | Sum of intensities | | `voxel_count` | Number of voxels in region | | `z_filtered_mean` | Mean after removing outliers (|z| > 3) | | `z_filtered_std` | Std after removing outliers | | `skewness` | Distribution skewness | | `kurtosis` | Distribution kurtosis | ## Available Statistics | Statistic | Description | Edge Case Behavior | |-----------|-------------|-------------------| | `mean` | Arithmetic mean | NaN for empty regions | | `std` | Standard deviation | 0 for constant values | | `median` | 50th percentile | NaN for empty regions | | `mad_median` | Median absolute deviation | Robust alternative to std | | `min` / `max` | Extreme values | NaN for empty regions | | `range` | Max - Min | 0 for constant values | | `volume` | Sum of all values | Region-specific metric | | `voxel_count` | Number of non-zero voxels | Proxy for region size | | `z_filtered_mean` | Mean excluding |z| > 3 outliers | Robust to outliers | | `z_filtered_std` | Std excluding outliers | Robust variance estimate | | `iqr_filtered_mean` | Mean excluding IQR outliers | Alternative robust mean | | `robust_mean` | MAD-based filtered mean | Highly robust | | `skewness` | Asymmetry of distribution | Higher moments | | `kurtosis` | Tail heaviness | Outlier sensitivity | ## Python API Use the core parcellation engine directly: ```python from parcellate import VolumetricParcellator import nibabel as nib import pandas as pd # Load atlas atlas_img = nib.load("atlas.nii.gz") lut = pd.read_csv("atlas.tsv", sep="\t") # Initialize parcellator parcellator = VolumetricParcellator( atlas_img=atlas_img, lut=lut, background_label=0, resampling_target="data" # Resample to scalar map space ) # Parcellate a scalar map scalar_img = nib.load("gm_map.nii.gz") parcellator.fit(scalar_img) stats = parcellator.transform(scalar_img) print(stats.head()) ``` ### Custom Statistics Define custom aggregation functions: ```python def range_iqr(values): """Interquartile range.""" q75, q25 = np.percentile(values[~np.isnan(values)], [75, 25]) return q75 - q25 parcellator = VolumetricParcellator( atlas_img=atlas_img, stat_functions={"range_iqr": range_iqr} ) ``` ## Development ### Setup ```bash # Clone repository git clone https://github.com/GalKepler/parcellate.git cd parcellate # Install development dependencies make install # Run tests make test # Run code quality checks make check ``` ### Running Tests ```bash # All tests uv run python -m pytest # Specific test file uv run python -m pytest tests/test_parcellator.py # With coverage uv run python -m pytest --cov=src/parcellate --cov-report=html ``` ### Contributing 1. Fork the repository 2. Create a feature branch (`git checkout -b feature/amazing-feature`) 3. Make your changes 4. Run tests and quality checks (`make check && make test`) 5. Commit your changes (`git commit -m 'Add amazing feature'`) 6. Push to the branch (`git push origin feature/amazing-feature`) 7. Open a Pull Request ## Documentation Full documentation is available at [https://GalKepler.github.io/parcellate/](https://GalKepler.github.io/parcellate/) ## License This project is licensed under the MIT License - see the LICENSE file for details. ## Citation If you use this tool in your research, please cite: ```bibtex @software{parcellate, author = {Kepler, Gal}, title = {parcellate: Atlas-based parcellation of neuroimaging data}, url = {https://github.com/GalKepler/parcellate}, year = {2024} } ``` ## Acknowledgments Repository initiated with [fpgmaas/cookiecutter-uv](https://github.com/fpgmaas/cookiecutter-uv).
text/markdown
null
Gal Kepler <galkepler@gmail.com>
null
null
null
python
[ "Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Py...
[]
null
null
<4.0,>=3.9
[]
[]
[]
[ "nibabel>=5.3.2", "nilearn>=0.12.1", "numpy>=2.0.2", "pandas>=2.3.3", "scipy>=1.10.0", "tomli>=2.3.0", "python-dotenv>=1.0.0; extra == \"dotenv\"" ]
[]
[]
[]
[ "Homepage, https://GalKepler.github.io/parcellate/", "Repository, https://github.com/GalKepler/parcellate", "Documentation, https://GalKepler.github.io/parcellate/" ]
uv/0.9.28 {"installer":{"name":"uv","version":"0.9.28","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}
2026-02-19T12:02:16.176995
parcellate-0.2.0.tar.gz
325,825
4e/a4/89efdffaefd429623752dc79449e3ffdff18fa71482db52420aaf8e290bb/parcellate-0.2.0.tar.gz
source
sdist
null
false
16c2b6a0558d7eb34d36ea36eb06bf1d
d07a9837d580c0165cdd4aaea277ed58379426cf37d8f1561f1157add5b3950b
4ea489efdffaefd429623752dc79449e3ffdff18fa71482db52420aaf8e290bb
null
[ "LICENSE" ]
351
2.4
krystal-cloud
0.1.2
Python SDK for the Krystal Cloud API
# krystal-cloud Python SDK for the [Krystal Cloud API](https://cloud-api.krystal.app). ## Installation ```bash pip install krystal-cloud ``` ## Quick Start ```python from krystal_cloud import KrystalCloud client = KrystalCloud(api_key="your-api-key") # Get supported chains chains = client.get_chains() # Get wallet balances balances = client.get_balances("0xYourWallet", chain_ids=[1, 137]) # Get pools on Ethereum pools = client.get_pools(chain_id=1, limit=10, sort_by=1) # Get pool detail pool = client.get_pool(1, "0xPoolAddress") # Get positions positions = client.get_positions("0xYourWallet", chain_ids=[1]) # Get strategies strategies = client.get_strategies("0xYourWallet") ``` ## Async Usage ```python import asyncio from krystal_cloud import AsyncKrystalCloud async def main(): async with AsyncKrystalCloud(api_key="your-api-key") as client: chains = await client.get_chains() balances = await client.get_balances("0xYourWallet") asyncio.run(main()) ``` ## Error Handling ```python from krystal_cloud import KrystalCloud, AuthError, InsufficientCreditsError, KrystalAPIError client = KrystalCloud(api_key="your-api-key") try: data = client.get_chains() except AuthError: print("Invalid API key") except InsufficientCreditsError: print("Need more credits") except KrystalAPIError as e: print(f"API error {e.status_code}: {e}") ``` ## API Methods | Method | Description | |--------|-------------| | `get_balances(wallet, ...)` | Wallet token balances | | `get_chains()` | Supported chains | | `get_chain(chain_id)` | Chain stats | | `get_pools(...)` | Pool list with filters | | `get_pool(chain_id, pool_address)` | Pool detail | | `get_pool_historical(chain_id, pool_address)` | Pool historical data | | `get_pool_ticks(chain_id, pool_address)` | Pool ticks | | `get_pool_transactions(chain_id, pool_address)` | Pool transactions | | `get_positions(wallet, ...)` | User positions | | `get_position(chain_id, position_id)` | Position detail | | `get_position_performance(chain_id, position_id)` | Position performance | | `get_position_transactions(chain_id, position_id)` | Position transactions | | `get_protocols()` | Supported protocols | | `get_strategies(wallet, ...)` | Strategies by wallet | | `get_strategy_positions(strategy_id)` | Strategy positions | ## API Documentation Full API documentation: https://cloud-api.krystal.app/swagger/index.html ## License MIT
text/markdown
null
null
null
null
null
null
[]
[]
null
null
>=3.9
[]
[]
[]
[ "httpx>=0.24.0", "pytest; extra == \"dev\"", "pytest-asyncio; extra == \"dev\"", "respx; extra == \"dev\"" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.9.6
2026-02-19T12:01:54.885165
krystal_cloud-0.1.2.tar.gz
3,953
b9/6c/cc50ef4e2029834cf0d6dc5adb37d07657af535fb343d23a25abdc159502/krystal_cloud-0.1.2.tar.gz
source
sdist
null
false
d5fbc9a11663cdea2543c0d569e90bf1
6cefb0671a39b229b7ba094156c07e65de761fc342da19639b60a06554631eec
b96ccc50ef4e2029834cf0d6dc5adb37d07657af535fb343d23a25abdc159502
MIT
[]
233
2.4
rrtftree
1.0.1
Terraform Project Tree Generator CLI Tool
# 🚀 rtftree --- Terraform Project Tree Generator CLI > A professional-grade CLI tool to generate clean, structured, and > shareable Terraform project trees --- with smart exclude support, > colored output, and Markdown export. ------------------------------------------------------------------------ ![Python](https://img.shields.io/badge/python-3.8+-blue.svg) ![License](https://img.shields.io/badge/license-MIT-green.svg) ![CLI](https://img.shields.io/badge/type-CLI-orange.svg) ![Terraform](https://img.shields.io/badge/terraform-supported-purple.svg) ------------------------------------------------------------------------ ## 📌 Why rtftree? Terraform projects often include: - `.terraform/` - `terraform.tfstate` - `.terraform.lock.hcl` - `.git/` - Provider binaries - Deeply nested modules Sharing structure manually becomes messy and unreadable. 🔥 **rtftree solves this problem** by generating a clean, structured tree view of your infrastructure project --- ready for documentation, sharing, and audits. ------------------------------------------------------------------------ # ✨ Features ✅ Beautiful tree-style output\ ✅ 🎨 Colored CLI output\ ✅ Optional file content preview\ ✅ Smart exclude support (like `.gitignore`)\ ✅ Wildcard pattern support (`*.exe`, `.terraform*`)\ ✅ Exclude via file (`--exclude-file`)\ ✅ Markdown export mode (`--markdown`)\ ✅ Output to file (`-o`)\ ✅ Lightweight & Fast\ ✅ Installable as a CLI tool ------------------------------------------------------------------------ # 📦 Installation ## 🔹 Local Install (Development Mode) ``` bash python -m pip install -e . ``` ## 🔹 Standard Install ``` bash python -m pip install . ``` ## 🔹 After PyPI Publish (Global Install) ``` bash pip install rtftree ``` ------------------------------------------------------------------------ # 🚀 Usage ## Basic Usage ``` bash rtftree <project-folder> ``` Example: ``` bash rtftree . ``` ------------------------------------------------------------------------ ## 📁 Structure Only (No File Content) ``` bash rtftree . --no-content ``` ------------------------------------------------------------------------ ## 🚫 Exclude Files & Folders ### Direct Patterns ``` bash rtftree . --exclude .terraform .git terraform.tfstate *.exe ``` ### Using Exclude File ``` bash rtftree . --exclude-file exclude.txt ``` Example `exclude.txt`: .terraform .git terraform.tfstate *.exe .terraform.lock.hcl Supported: - Exact file names - Folder names - Wildcards ------------------------------------------------------------------------ ## 💾 Save Output to File ``` bash rtftree . -o infra_tree.txt ``` ------------------------------------------------------------------------ ## 📝 Markdown Export Mode Generate Markdown-ready structure: ``` bash rtftree . --markdown -o structure.md ``` Perfect for: - GitHub documentation - Wiki pages - Confluence - Client documentation ------------------------------------------------------------------------ # 🖥 Example Output 📁 Terraform Project: infra ├── 📁 modules │ ├── 📄 main.tf │ │ resource "azurerm_resource_group" "rg" { │ │ name = "example" │ │ location = "East US" │ │ } │ └── 📄 variables.tf │ variable "location" { │ type = string │ } └── 📄 provider.tf ------------------------------------------------------------------------ # ⚙️ CLI Options Option Description ------------------ ------------------------------------ `--no-content` Show only folder/file structure `--exclude` Space-separated patterns to ignore `--exclude-file` Load exclude patterns from file `--markdown` Export output in Markdown format `-o`, `--output` Write output to file ------------------------------------------------------------------------ # 🏗 Project Structure rtftree/ │ ├── rtftree/ │ ├── __init__.py │ └── cli.py │ ├── pyproject.toml └── README.md ------------------------------------------------------------------------ # 🔥 DevOps Use Cases - Share Terraform structure in tickets - Infrastructure documentation - CI/CD pipeline documentation - Client infrastructure overview - Audit reporting - Pre-deployment reviews ------------------------------------------------------------------------ # 🌍 PyPI Publishing Once published to PyPI, anyone can install globally: ``` bash pip install rtftree ``` This makes rtftree a globally accessible DevOps utility tool. ------------------------------------------------------------------------ # 🧠 Roadmap - `.treeignore` auto-detection - `--max-depth` option - Terraform-only mode (`*.tf`) - JSON export - GitHub Action integration - Auto documentation mode ------------------------------------------------------------------------ # 👨‍💻 Author **Ritesh Sharma**\ DevOps \| Azure \| Terraform \| Kubernetes ------------------------------------------------------------------------ # 📄 License MIT License ------------------------------------------------------------------------ # ⭐ Support If you find this tool useful: ⭐ Star the repository\ 🚀 Share with DevOps community\ 🛠 Contribute improvements ------------------------------------------------------------------------ > Built with ❤️ for DevOps Engineers
text/markdown
Ritesh Sharma
null
null
null
MIT
null
[]
[]
null
null
>=3.8
[]
[]
[]
[ "rich>=13.0.0" ]
[]
[]
[]
[ "Homepage, https://github.com/riteshatri/rtftree" ]
twine/6.2.0 CPython/3.11.9
2026-02-19T12:01:09.507950
rrtftree-1.0.1.tar.gz
5,498
f9/69/f32dd933233cc6584945dfadbb400baa9dd5edeca03b01d16402066e1d78/rrtftree-1.0.1.tar.gz
source
sdist
null
false
58e1ce3cd0094c6b98eb5246c0179061
a1dccb052bd0138129df0da1f33efbe74525deed7f67001be264e6a725080755
f969f32dd933233cc6584945dfadbb400baa9dd5edeca03b01d16402066e1d78
null
[ "LICENSE" ]
178
2.1
coati-payroll
1.7.3
A jurisdiction-agnostic payroll calculation engine.
![Logo](https://raw.githubusercontent.com/bmosoluciones/coati-payroll/refs/heads/main/coati_payroll/static/logo/Icono-small.png) # Coati Payroll ![PyPI - License](https://img.shields.io/badge/License-Apache%202.0-brightgreen?logo=apache&logoColor=white) ![Python 3.11+](https://img.shields.io/badge/python-3.11+-blue.svg?logo=python&logoColor=white) [![Unit Testing](https://github.com/williamjmorenor/coati-payroll/actions/workflows/python-package.yml/badge.svg?branch=main)](https://github.com/williamjmorenor/coati-payroll/actions/workflows/python-package.yml) [![codecov](https://codecov.io/github/bmosoluciones/coati-payroll/graph/badge.svg?token=gT6jVct2KT)](https://codecov.io/github/bmosoluciones/coati-payroll) [![Code style: black](https://img.shields.io/badge/Python%20code%20style-black-000000.svg)](https://github.com/psf/black) [![Code style: Prettier](https://img.shields.io/badge/HTML%20code%20style-prettier-ff69b4.svg)](https://github.com/prettier/prettier) [![Ruff](https://img.shields.io/endpoint?url=https://raw.githubusercontent.com/astral-sh/ruff/main/assets/badge/v2.json)](https://github.com/astral-sh/ruff) A jurisdiction-agnostic payroll calculation engine developed by BMO Soluciones, S.A. ![ScreenShot](https://raw.githubusercontent.com/bmosoluciones/coati-payroll/refs/heads/main/docs/CoatiPayrollSystem.png) ----------- [![ScreenShot](https://via.placeholder.com/200x150.png?text=Dashboard)](https://via.placeholder.com/800x600.png?text=Dashboard) [![ScreenShot](https://via.placeholder.com/200x150.png?text=Employee+Management)](https://via.placeholder.com/800x600.png?text=Employee+Management) [![ScreenShot](https://via.placeholder.com/200x150.png?text=Payroll+Processing)](https://via.placeholder.com/800x600.png?text=Payroll+Processing) [![ScreenShot](https://via.placeholder.com/200x150.png?text=Reports)](https://via.placeholder.com/800x600.png?text=Reports) ----------- **Coati Payroll** is a flexible and extensible payroll calculation engine, completely jurisdiction-agnostic. The system is designed so that organizations and implementers can define their own payroll rules through configuration, without the need to modify the source code. The engine does not incorporate hardcoded legal rules. All earnings, deductions, benefits, taxes, and other payroll concepts exist only if the implementer configures them. > **Important**: This project is governed by a [Social Contract](SOCIAL_CONTRACT.md) that clearly defines its scope, purpose, and limitations. Please read it before using the system in production. ## Disclaimer and scope ``` The calculation engine may not be suitable for all possible use cases. The development team is happy to collaborate with interested parties to expand the engine's capabilities so it can cover as many use cases as possible. This project is offered under an open-source license, as-is, without warranties of fitness for any particular purpose. ``` ## Key Features - **Jurisdiction-Agnostic**: No hardcoded legal rules; all payroll logic is defined through configuration - **Configurable Calculation Engine**: Payroll processing with fully configurable formulas and schemas - **Flexible Calculation Rules**: Rule system that allows implementing any payroll logic through configuration - **Multi-company**: Manage payrolls for multiple companies or entities from a single system - **Employee Management**: Complete registration of personal, work, and salary information - **Custom Fields**: Extend employee information with custom fields - **Configurable Earnings**: Define any type of additional income (bonuses, commissions, overtime, etc.) - **Priority-Based Deductions**: Configure deductions in priority order according to your needs - **Employer Benefits**: Configure benefits and employer contributions as required - **Planilla Cloning (Web UI)**: Duplicate an existing payroll template from the planilla list, including perceptions, deductions, and benefits - **Loans and Advances**: Loan control with automatic installment deduction - **Multi-currency**: Support for multiple currencies with exchange rates - **Background Processing**: Queue system for large payrolls with Dramatiq+Redis - **Vacation Management**: Complete module for vacation accrual, usage, and audit with configurable policies - **Role-Based Access Control (RBAC)**: Permission system with Admin, HR, and Audit roles - **Reporting System**: Custom reports with role-based permissions and execution audit - **Internationalization**: Multi-language support with interface and content translation - **Periodicidad Compatibility**: Salary calculation accepts both Spanish and English periodicity terms (`mensual`/`monthly`, `quincenal`/`biweekly`) ## Quick Installation ### Requirements - Python 3.11 or higher - pip (Python package manager) ### Steps 1. **Clone the repository** ```bash git clone https://github.com/williamjmorenor/coati-payroll.git cd coati-payroll ``` 2. **Create and activate virtual environment** ```bash python -m venv venv source venv/bin/activate # Linux/macOS # or venv\Scripts\activate # Windows ``` 3. **Install dependencies** ```bash pip install -r requirements.txt ``` 4. **Run the application** ```bash python app.py ``` 5. **Access the system** Open your browser at `http://localhost:5000` **Default credentials:** - User: `coati-admin` - Password: `coati-admin` > **Important**: Change the default credentials in production environments. ### Docker Installation Coati Payroll supports three deployment architectures with Docker: #### Architecture Options 1. **Option 1: Single container without queue processing** - Simplest setup for development or small deployments 2. **Option 2: Single container with background processing** - All-in-one deployment with app + worker in same container 3. **Option 3: Separate containers (Production)** - Web and worker in separate containers for scalability --- #### Option 1: Single Container Without Queue Processing For development or when background processing is not needed: ```bash # Build the image docker build -t coati-payroll:latest . # Run without queue processing docker run -d -p 5000:5000 \ -e FLASK_ENV=development \ -e QUEUE_ENABLED=0 \ --name coati-payroll \ coati-payroll:latest ``` **Behavior**: All payroll calculations execute synchronously. No Redis required. --- #### Option 2: Single Container With Background Processing (All-in-one) For small to medium deployments where separating web/worker isn't necessary: ```bash # Start Redis docker run -d -p 6379:6379 --name redis redis:alpine # Start Coati Payroll with worker in same container docker run -d -p 5000:5000 \ -e FLASK_ENV=production \ -e DATABASE_URL="postgresql://user:password@host:5432/coati_payroll" \ -e SECRET_KEY="your-secret-key-here" \ -e ADMIN_USER="admin" \ -e ADMIN_PASSWORD="secure-password" \ -e QUEUE_ENABLED=1 \ -e PROCESS_ROLE=all \ -e REDIS_URL="redis://redis:6379/0" \ -e BACKGROUND_PAYROLL_THRESHOLD=100 \ -e DRAMATIQ_WORKER_THREADS=8 \ -e DRAMATIQ_WORKER_PROCESSES=2 \ --link redis:redis \ --name coati-payroll \ coati-payroll:latest ``` **Behavior**: - Web app serves requests - Dramatiq worker runs in background in same container - Suitable for deployments with moderate load - Default behavior when `PROCESS_ROLE` is not specified **Verify worker started**: ```bash docker logs coati-payroll # Should show: "[entrypoint] Starting Dramatiq worker in background (all-in-one mode, threads=8, processes=2)" ``` --- #### Option 3: Separate Containers (Recommended for Production) For production deployments requiring scalability and fault isolation: **Using Docker Compose** (recommended): A production-ready `docker-compose.yml` is provided in the repository with: - **Nginx reverse proxy** (serves static files and handles HTTPS) - PostgreSQL database (with MySQL as commented alternative) - Redis for queue and cache - Dedicated worker container - Web application container (WSGI server) - Certbot for Let's Encrypt SSL certificates (optional) - Health checks and proper service dependencies **Quick start**: ```bash # 1. Copy and customize environment variables cp .env.example .env # Edit .env and set secure passwords and secrets! # 2. Start all services docker-compose up -d # 3. View logs docker-compose logs -f # 4. Access the application # HTTP: http://localhost or http://your-server-ip # HTTPS: https://your-domain.com (after configuring SSL) ``` **Static files are served by nginx**, not the WSGI server. This provides: - Faster delivery of CSS, JavaScript, and images - Reduced load on the application server - Better caching and compression **HTTPS setup with Let's Encrypt**: For production deployments with SSL certificates: ```bash # Run the Let's Encrypt initialization script chmod +x nginx/init-letsencrypt.sh ./nginx/init-letsencrypt.sh your-domain.com your-email@example.com # Edit nginx/nginx.conf and uncomment the HTTPS server block # Uncomment the certbot service in docker-compose.yml # Restart services docker-compose restart nginx ``` See `nginx/README.md` for detailed HTTPS configuration instructions. **Scale workers** (if you need more processing capacity): ```bash docker-compose up -d --scale worker=3 ``` **Using MySQL instead of PostgreSQL**: Edit `docker-compose.yml` and: 1. Comment out the `postgres` service 2. Uncomment the `mysql` service 3. Update `DATABASE_URL` in `web` and `worker` services 4. Update `depends_on` to reference `mysql` instead of `postgres` **Or using individual containers**: ```bash # Build image docker build -t coati-payroll:latest . # Start infrastructure docker run -d --name redis redis:alpine docker run -d --name postgres -e POSTGRES_PASSWORD=changeme postgres:15-alpine # Start web container (no worker) docker run -d -p 5000:5000 \ --name coati-web \ -e PROCESS_ROLE=web \ -e QUEUE_ENABLED=1 \ -e REDIS_URL=redis://redis:6379/0 \ -e DATABASE_URL=postgresql://postgres:changeme@postgres:5432/postgres \ --link redis:redis \ --link postgres:postgres \ coati-payroll:latest # Start worker container (dedicated) docker run -d \ --name coati-worker \ -e PROCESS_ROLE=worker \ -e QUEUE_ENABLED=1 \ -e REDIS_URL=redis://redis:6379/0 \ -e DATABASE_URL=postgresql://postgres:changeme@postgres:5432/postgres \ -e DRAMATIQ_WORKER_THREADS=8 \ -e DRAMATIQ_WORKER_PROCESSES=2 \ --link redis:redis \ --link postgres:postgres \ coati-payroll:latest ``` **Benefits of Option 3**: - Scale web and worker independently - Isolate failures (worker crash doesn't affect web) - Optimize resource allocation per component - Standard production architecture pattern **Verify services**: ```bash # Check web container docker logs coati-web # Should show: "[entrypoint] Starting app in web-only mode (no worker)" # Check worker container docker logs coati-worker # Should show: "[entrypoint] Starting Dramatiq worker (dedicated mode, threads=8, processes=2)" ``` --- #### Environment Variables Reference | Variable | Description | Default | Required for | |----------|-------------|---------|--------------| | `PROCESS_ROLE` | Container role: `web`, `worker`, or `all` | `all` | Options 2, 3 | | `QUEUE_ENABLED` | Enable queue system | `1` | Options 2, 3 | | `REDIS_URL` | Redis connection string | - | Options 2, 3 | | `BACKGROUND_PAYROLL_THRESHOLD` | Min employees for background processing | `100` | Options 2, 3 | | `DRAMATIQ_WORKER_THREADS` | Worker threads per process | `8` | Options 2, 3 | | `DRAMATIQ_WORKER_PROCESSES` | Worker processes | `2` | Options 2, 3 | **Access the system**: Open your browser at `http://localhost:5000` ## Documentation Complete documentation is available in the `docs/` directory and can be generated with MkDocs: ```bash # Install documentation dependencies pip install -r docs.txt # Serve documentation locally mkdocs serve # Generate static documentation mkdocs build ``` ### Documentation Contents - **[Quick Start Guide](docs/guia/inicio-rapido.md)**: 15 minutes from installation to your first payroll - ideal for evaluating the system - **Installation Guide**: Requirements, installation, and initial configuration - **User Guide**: Users, companies, currencies, employees, custom fields, payroll concepts, calculation rules, loans, vacations, accounting configuration - **Complete Tutorial**: Step by step to configure and run a payroll with all components - **Advanced Features**: - Queue system and background processing - Database compatibility (SQLite, PostgreSQL, MySQL/MariaDB) - Role-based access control (RBAC) - Vacation management with configurable policies - Custom reporting system - Internationalization and translation - **Reference**: Glossary, frequently asked questions, exchange rate import ## Architecture ``` coati/ ├── app.py # Application entry point ├── coati_payroll/ # Main module │ ├── __init__.py # Flask application factory │ ├── model.py # Database models (SQLAlchemy) │ ├── nomina_engine/ # Payroll calculation engine (refactored) │ │ ├── __init__.py │ │ ├── engine.py # Main orchestrator │ │ ├── domain/ # Domain models │ │ │ ├── payroll_context.py │ │ │ ├── employee_calculation.py │ │ │ └── calculation_items.py │ │ ├── validators/ # Validations │ │ │ ├── base_validator.py │ │ │ ├── planilla_validator.py │ │ │ ├── employee_validator.py │ │ │ ├── period_validator.py │ │ │ └── currency_validator.py │ │ ├── calculators/ # Calculations │ │ │ ├── salary_calculator.py │ │ │ ├── concept_calculator.py │ │ │ ├── perception_calculator.py │ │ │ ├── deduction_calculator.py │ │ │ ├── benefit_calculator.py │ │ │ └── exchange_rate_calculator.py │ │ ├── processors/ # Specific processors │ │ │ ├── loan_processor.py │ │ │ ├── accumulation_processor.py │ │ │ ├── vacation_processor.py │ │ │ ├── novelty_processor.py │ │ │ └── accounting_processor.py │ │ ├── repositories/ # Data access │ │ │ ├── base_repository.py │ │ │ ├── planilla_repository.py │ │ │ ├── employee_repository.py │ │ │ ├── acumulado_repository.py │ │ │ ├── novelty_repository.py │ │ │ ├── exchange_rate_repository.py │ │ │ └── config_repository.py │ │ ├── services/ # Business services │ │ │ ├── payroll_execution_service.py │ │ │ └── employee_processing_service.py │ │ └── results/ # Results and DTOs │ │ ├── payroll_result.py │ │ ├── validation_result.py │ │ └── error_result.py │ ├── formula_engine/ # Formula engine (refactored) │ │ ├── __init__.py │ │ ├── engine.py # Main orchestrator │ │ ├── exceptions.py # Custom exceptions │ │ ├── data_sources.py # Available data sources │ │ ├── novelty_codes.py # Novelty codes │ │ ├── ast/ # Expression evaluation (Visitor pattern) │ │ │ ├── ast_visitor.py │ │ │ ├── expression_evaluator.py │ │ │ ├── safe_operators.py │ │ │ └── type_converter.py │ │ ├── validation/ # Validations │ │ │ ├── schema_validator.py │ │ │ ├── tax_table_validator.py │ │ │ └── security_validator.py │ │ ├── steps/ # Step types (Strategy pattern) │ │ │ ├── base_step.py │ │ │ ├── calculation_step.py │ │ │ ├── conditional_step.py │ │ │ ├── tax_lookup_step.py │ │ │ ├── assignment_step.py │ │ │ └── step_factory.py │ │ ├── tables/ # Tax tables │ │ │ ├── tax_table.py │ │ │ ├── bracket_calculator.py │ │ │ └── table_lookup.py │ │ ├── execution/ # Execution context │ │ │ ├── execution_context.py │ │ │ ├── step_executor.py │ │ │ └── variable_store.py │ │ └── results/ # Results │ │ └── execution_result.py │ ├── formula_engine_examples.py # Schema examples │ ├── vacation_service.py # Vacation management service │ ├── rbac.py # Role-based access control │ ├── report_engine.py # Reporting engine │ ├── forms.py # WTForms forms │ ├── cli.py # Command-line interface (payrollctl) │ ├── queue/ # Queue system (Dramatiq+Redis) │ │ ├── driver.py │ │ ├── selector.py │ │ ├── tasks.py │ │ └── drivers/ │ ├── vistas/ # Views/Controllers (Blueprints) │ │ ├── planilla/ # Payroll module │ │ └── [other modules] │ ├── templates/ # HTML templates (Jinja2) │ ├── translations/ # Translation files (i18n) │ └── static/ # Static files ├── docs/ # MkDocs documentation ├── requirements.txt # Production dependencies ├── development.txt # Development dependencies └── docs.txt # Documentation dependencies ``` ## Configuration ### Command Line Interface (CLI) The system includes the `payrollctl` tool for common administrative tasks. You can also use `flask` for built-in commands. **System Operations:** ```bash # View system status payrollctl system status # Run system checks payrollctl system check # View system information payrollctl system info # View environment variables payrollctl system env ``` **Database Management:** ```bash # View database status payrollctl database status # Initialize database and create admin user payrollctl database init # Load initial data (currencies, concepts, etc.) payrollctl database seed # Create database backup using native tools # SQLite: Copy file | PostgreSQL: pg_dump | MySQL: mysqldump payrollctl database backup -o backup_$(date +%Y%m%d).sql # Restore database from backup payrollctl database restore backup.db # Database migration (requires flask-migrate) payrollctl database migrate payrollctl database upgrade # Drop all tables (CAUTION!) payrollctl database drop ``` **User Management:** ```bash # List all users payrollctl users list # Create a new user payrollctl users create # Disable a user payrollctl users disable username # Reset password payrollctl users reset-password username # Create or update admin user (disables other admins) payrollctl users set-admin ``` **Cache Management:** ```bash # Clear application caches payrollctl cache clear # Warm up caches payrollctl cache warm # View cache status payrollctl cache status ``` **Maintenance Tasks:** ```bash # Clean up expired sessions payrollctl maintenance cleanup-sessions # Clean up temporary files payrollctl maintenance cleanup-temp # Run pending background jobs payrollctl maintenance run-jobs ``` **Diagnostics and Debugging:** ```bash # View application configuration payrollctl debug config # List all application routes payrollctl debug routes ``` **Note**: All commands also work with `flask` (e.g., `flask system status`). **Automated Backups**: To configure automatic daily backups with systemd timers, see [Automated Backups Guide](docs/automated-backups.md). ### Environment Variables | Variable | Description | Default Value | |----------|-------------|---------------| | `DATABASE_URL` | Database connection URI | Local SQLite | | `SECRET_KEY` | Secret key for sessions | Auto-generated | | `ADMIN_USER` | Initial admin user | `coati-admin` | | `ADMIN_PASSWORD` | Admin password | `coati-admin` | | `PORT` | Application port | `5000` | | `SESSION_REDIS_URL` | Redis URL for sessions | None (uses SQLAlchemy) | | `REDIS_URL` | Redis URL for queue system | None (background disabled) | | `QUEUE_ENABLED` | Enable queue system | `1` | | `BACKGROUND_PAYROLL_THRESHOLD` | Employee threshold for background processing | `100` | ### Database The system supports: - **SQLite**: For development and testing (default) - **PostgreSQL**: Recommended for production - **MySQL/MariaDB**: Production alternative The system is designed to be **database engine agnostic**. For more details on compatibility and configuration, see the [Database Compatibility Guide](docs/database-compatibility.md). ### Queue System For long-running operations, the system includes a **background process queue system**: - **Dramatiq + Redis**: Required for background processing - **Automatic degradation**: If Redis is unavailable, background processing is disabled and payrolls execute synchronously - **Parallel processing**: Large payrolls (above threshold) are automatically processed in the background when queue is enabled - **Real-time feedback**: Task progress tracking For more information, see the [Queue System Documentation](docs/queue_system.md) and [Background Payroll Processing](docs/background-payroll-processing.md). ## Workflow ```mermaid graph LR A[Configure Currencies] --> B[Create Earnings/Deductions/Benefits] B --> C[Register Employees] C --> D[Create Payroll] D --> E[Assign Components] E --> F[Execute Payroll] F --> G[Review and Approve] G --> H[Apply Payroll] ``` ## Payroll Calculation The payroll engine processes in this order: 1. **Base Salary**: Salary defined for the employee according to the payroll period 2. **Earnings**: Added to base salary → Gross Salary 3. **Deductions**: Subtracted in priority order → Net Salary 4. **Benefits**: Calculated as employer costs (do not affect net salary) ### Illustrative Calculation Example > **Important Note**: This is an illustrative example with generic values and concepts. Concept names, percentages, and specific calculations **must be configured by the implementer** according to the laws and policies of their jurisdiction. The engine **does not include predefined legal rules**. ``` Base Salary: $ 10,000.00 + Earning A: $ 500.00 + Earning B: $ 300.00 = GROSS SALARY: $ 10,800.00 - Deduction A (X%): $ 756.00 - Deduction B (Y%): $ 540.00 - Deduction C: $ 200.00 = NET SALARY: $ 9,304.00 Employer Benefits (Company Costs): + Benefit A (W%): $ 2,160.00 + Benefit B (Z%): $ 216.00 + Benefit C (P%): $ 899.64 + Benefit D (P%): $ 899.64 + Benefit E (P%): $ 899.64 = TOTAL COMPANY COST: $ 15,178.92 ``` **How to configure these concepts?** All concepts, percentages, and calculation rules are defined through: - **Configurable earnings**: Define any type of additional income - **Priority-based deductions**: Configure the order and formula for each deduction - **Employer benefits**: Configure contributions according to your jurisdiction - **Calculation rules**: Use the rule engine to implement complex logic (brackets, caps, exemptions, etc.) See the [complete documentation](docs/) to learn how to configure your payroll system. ## Development ### Install development dependencies ```bash pip install -r development.txt ``` ### Database Structure The main models are: **System Configuration:** - `Usuario`: System users with roles (Admin, HR, Audit) - `Empresa`: Companies or entities that hire employees - `Moneda`: System currencies - `TipoCambio`: Exchange rates between currencies - `ConfiguracionGlobal`: Global system configuration **Personnel Management:** - `Empleado`: Employee master record - `CampoPersonalizado`: Custom fields for employees - `HistorialSalario`: Salary change history **Payroll:** - `Percepcion`: Income concepts - `Deduccion`: Deduction concepts - `Prestacion`: Employer contributions - `ReglaCalculo`: Calculation rules with configurable schemas - `TipoPlanilla`: Payroll types (monthly, biweekly, etc.) - `Planilla`: Payroll configuration - `Nomina`: Payroll execution - `NominaEmpleado`: Payroll detail per employee - `NominaDetalle`: Detail lines (earnings, deductions) - `NominaNovedad`: Payroll novelties - `ComprobanteContable`: Accounting vouchers **Loans:** - `Adelanto`: Employee loans and advances - `AdelantoAbono`: Loan payments **Vacations:** - `VacationPolicy`: Configurable vacation policies - `VacationAccount`: Vacation accounts per employee - `VacationLedger`: Vacation ledger (audit) - `VacationNovelty`: Vacation requests and novelties - `ConfiguracionVacaciones`: Vacation configuration (legacy) - `VacacionEmpleado`: Employee vacations (legacy) - `PrestacionAcumulada`: Accumulated benefits - `CargaInicialPrestacion`: Initial benefit load **Reports:** - `Report`: Custom report definitions - `ReportRole`: Report permissions by role - `ReportExecution`: Report execution history - `ReportAudit`: Report audit ## Social Contract and Responsibilities This project is governed by a [Social Contract](SOCIAL_CONTRACT.md) that clearly establishes: ### Project Scope - **Jurisdiction-agnostic engine**: Does not include and will not include hardcoded legal rules - **Strict separation**: Between calculation engine, rule configuration, and payroll orchestration - **Predictable and reproducible calculation**: Calculations are deterministic and auditable - **Extensible by configuration**: Any legal change is implemented through configuration, not code ### Default Functionality The engine, by default, **only** calculates: 1. Employee base salary according to the defined period 2. Salary advance installments when they exist All other concepts (earnings, deductions, benefits, taxes, caps, brackets, exemptions) exist only if the implementer configures them. ### Implementer Responsibility Correct use of the engine requires that the implementer: - Has knowledge of how payroll is calculated in their jurisdiction - Understands the applicable legal framework - Is capable of manually calculating a complete payroll - Compares manual results with system results - Identifies and corrects configuration errors ### Warranties and Limitations This software is distributed under the Apache 2.0 License **"AS IS"**: - ✅ **Promises**: Predictable, reproducible, and auditable calculations - ✅ **Promises**: Remain jurisdiction-agnostic - ✅ **Promises**: Separation between engine and configuration - ❌ **Does not guarantee**: Regulatory compliance in any jurisdiction - ❌ **Does not guarantee**: Correct results without appropriate configuration - ❌ **Does not replace**: Professional knowledge or legal advice **For more details, read the complete [Social Contract](SOCIAL_CONTRACT.md) before using this system in production.** ## Support To report issues or request features, please open an [Issue on GitHub](https://github.com/williamjmorenor/coati-payroll/issues). ## License ``` SPDX-License-Identifier: Apache-2.0 Copyright 2025 - 2026 BMO Soluciones, S.A. ``` This project is licensed under the **Apache License 2.0** - a permissive open-source license that allows free use, modification, and distribution (including for commercial purposes), as long as copyright and license notices are preserved. It also includes a patent grant to protect users from patent claims but terminates rights if you file such claims. You may combine Apache-licensed code with proprietary software, but you cannot use Apache trademarks or logos without permission, and you must provide proper attribution to the original authors. For more details, see the [LICENSE](LICENSE) file. ## Contributing Contributions are welcome. Please: 1. Fork the repository 2. Create a branch for your feature (`git checkout -b feature/new-feature`) 3. Commit your changes (`git commit -am 'Add new feature'`) 4. Push to the branch (`git push origin feature/new-feature`) 5. Open a Pull Request --- Made with ❤️ by [BMO Soluciones, S.A.](https://github.com/williamjmorenor)
text/markdown
null
null
null
null
null
null
[]
[]
null
null
>=3.11
[]
[]
[]
[ "alembic==1.18.1", "argon2-cffi==25.1.0", "babel==2.17.0", "configobj==5.0.9", "cryptography==46.0.3", "dramatiq[redis]==2.0.0", "email_validator==2.3.0", "flask==3.1.2", "flask-alembic==3.2.0", "flask-babel==4.0.0", "flask-caching==2.3.1", "flask-limiter==4.1.1", "flask-login==0.6.3", "fl...
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:01:00.458785
coati_payroll-1.7.3.tar.gz
484,476
f7/86/21cc154052f3d35af5ef375785786afae289fc7c3988a3e2f890c4a18bf3/coati_payroll-1.7.3.tar.gz
source
sdist
null
false
cc792d4a55b56feb5f63468905ffc8b6
e7746ab6e1d5efa7d383719b4041530ab8b6f8307077ef3378eb89a0c514c45d
f78621cc154052f3d35af5ef375785786afae289fc7c3988a3e2f890c4a18bf3
null
[]
248
2.4
free-energy
0.0.1
A professional-grade AI utility for automated data synchronization and backend management.
# Installation To install requirements: `python -m pip install requirements.txt` To save requirements: `python -m pip list --format=freeze --exclude-editable -f https://download.pytorch.org/whl/torch_stable.html > requirements.txt` * Note we use Python 3.9.4 for our experiments # Running the code For remaining experiments: Navigate to the corresponding directory, then execute: `python run.py -m` with the corresponding `config.yaml` file (which stores experiment configs). # License Consult License.md
text/markdown
null
AI Research Team <Ai-model@example.com>
null
null
null
automation, api-client, sync, tooling
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent" ]
[]
null
null
>=3.8
[]
[]
[]
[ "requests>=2.28.0", "urllib3>=1.26.0" ]
[]
[]
[]
[ "Homepage, https://github.com/ai/library", "Bug Tracker, https://github.com/ai/library/issues" ]
twine/6.2.0 CPython/3.14.3
2026-02-19T12:00:31.864698
free_energy-0.0.1.tar.gz
3,567
ed/83/53b9bd6513deb4e1eeea4f426a9987c903def6dfceb00373a98571f1d0b7/free_energy-0.0.1.tar.gz
source
sdist
null
false
5cf34fb727bf8db963e9dfc0ec800b50
5e8269364b3c7da92db744167843ee97d3e74d6eddbd9570aedb31ee6d75ecc8
ed8353b9bd6513deb4e1eeea4f426a9987c903def6dfceb00373a98571f1d0b7
null
[ "LICENSE.txt" ]
204
2.4
sahim-django-common
0.1.2
Django utilities and common reusable components
# sahim-django-common 📦 **Django utilities and common components** developed and maintained by [SahimCo](https://github.com/sahimco). This package is published publicly on [PyPI](https://pypi.org/project/sahim-django-common/) for use in any Django project. ## 📥 Installation Install from PyPI with pip: ```bash pip install sahim-django-common ``` Or add to your project's dependencies (e.g. `requirements.txt`): ``` sahim-django-common ``` ## 📋 Requirements - Python 3.10+ - Django 5.0+ ## 🚀 Usage Add the app to your Django project's `INSTALLED_APPS`: ```python INSTALLED_APPS = [ # ... "sahim_django_common", # ... ] ``` Or with the full app config: ```python INSTALLED_APPS = [ # ... "sahim_django_common.apps.SahimDjangoCommonConfig", # ... ] ``` After installation, follow the documentation for the specific utilities you need. Usage details will depend on which modules you use. *(Documentation for individual utilities will be added here or in the docs as the package grows.)* ## 📦 What's included This package provides reusable Django utilities used across SahimCo projects, such as: - 🔧 Helper functions and decorators - ⚡ Optional management commands and app hooks *(This section can be updated as new utilities are added.)* ## 📄 License This project is licensed under the MIT License — see the [LICENSE](LICENSE) file for details. ## 🔧 Development ### Setup Install the package in editable mode with all dev and test dependencies: ```bash pip install -r requirements-dev.txt ``` This installs the package (editable) plus pre-commit, pytest, and pytest-django. ### Pre-commit This project uses [pre-commit](https://pre-commit.com/) for code quality checks: ```bash # Install git hooks (run once) pre-commit install ``` Hooks run automatically on commit. To run manually: ```bash pre-commit run --all-files ``` ### Tests Run tests with pytest: ```bash pytest ``` ## 📤 Publishing releases (PyPI) Releases are published to PyPI automatically via GitHub Actions when you push a version tag. ### One-time setup: Trusted Publishing on PyPI 1. Log in to [PyPI](https://pypi.org) and open your project’s **Publishing** settings: - `https://pypi.org/manage/project/sahim-django-common/settings/publishing/` 1. Add a **Trusted Publisher** with: - **Owner:** `sahimco` (or your GitHub org/user) - **Repository name:** `sahim-django-common` - **Workflow name:** `release.yml` - **Environment name:** `pypi` 1. Create a `pypi` environment in GitHub: Repo → **Settings** → **Environments** → **New environment** → name it `pypi` (Optional: enable **Required reviewers** for extra safety.) ### Releasing a new version 1. Bump version in `pyproject.toml`. 1. Commit and push: ```bash git add pyproject.toml git commit -m "Release v0.1.1" git tag v0.1.1 git push origin main && git push origin v0.1.1 ``` 1. The workflow runs automatically and publishes to PyPI. ## 🤝 Contributing Contributions are welcome. Please open an issue or pull request on the [GitHub repository](https://github.com/sahimco/sahim-django-common). ______________________________________________________________________ **SahimCo** — [GitHub](https://github.com/sahimco) · [PyPI](https://pypi.org/project/sahim-django-common/)
text/markdown
null
SahimCo <sahimco.develop@gmail.com>
null
null
MIT License Copyright (c) 2026 SahimCo Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
django, utilities, common, reusable, sahimco
[ "Development Status :: 4 - Beta", "Environment :: Web Environment", "Framework :: Django", "Framework :: Django :: 5.0", "Framework :: Django :: 5.1", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python ::...
[]
null
null
>=3.10
[]
[]
[]
[ "Django<6.0,>=5.0", "pre-commit>=4.0; extra == \"dev\"", "pytest; extra == \"test\"", "pytest-django; extra == \"test\"" ]
[]
[]
[]
[ "Homepage, https://github.com/sahimco/sahim-django-common", "Documentation, https://github.com/sahimco/sahim-django-common#readme", "Repository, https://github.com/sahimco/sahim-django-common", "Bug Tracker, https://github.com/sahimco/sahim-django-common/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-19T12:00:29.940522
sahim_django_common-0.1.2.tar.gz
5,023
7b/be/da9c94dc91658901189489ff3d58d76b62ea7dc1a0f26628454c753031cc/sahim_django_common-0.1.2.tar.gz
source
sdist
null
false
062e73643c93007d222012120c787ede
0f29a25132f3e42c05fe18b51de5b68339ba1907653951ca64970576a2053726
7bbeda9c94dc91658901189489ff3d58d76b62ea7dc1a0f26628454c753031cc
null
[ "LICENSE" ]
234
2.4
penguin-metrics
0.0.2
Linux system telemetry service that sends data to MQTT, with Home Assistant integration via MQTT Discovery
# Penguin Metrics [![Tests](https://github.com/clusterm/penguin-metrics/actions/workflows/test.yml/badge.svg)](https://github.com/clusterm/penguin-metrics/actions/workflows/test.yml) Linux system telemetry service that sends data to MQTT, with Home Assistant integration via MQTT Discovery. **Why monitor your servers through Home Assistant?** - 📊 **All in one place** — see your servers alongside smart home devices in a single dashboard - 🪶 **Lightweight** — no need for heavy monitoring stacks like Prometheus + Grafana - 🔔 **Smart automations** — get notifications when disk space is low, CPU is overloaded, or a service goes down - 📈 **Beautiful visualization** — Home Assistant offers flexible cards, graphs, and history tracking - 🔌 **Zero configuration on HA side** — MQTT Discovery automatically creates all sensors and devices - 🏠 **Perfect for home servers** — simple setup, minimal resource usage, native HA integration ## Features ### Data Collection - **System Metrics**: CPU (overall and per-core), RAM, swap, load average, uptime, disk I/O (bytes + KiB/s rate), CPU frequency (MHz), process count, boot time - **Temperature**: Thermal zones and hwmon sensors (auto-discovery supported) - **Disk Space**: Total, used, free space and usage percentage (auto-discovery supported) - **Process Monitoring**: By name, regex pattern, PID, or pidfile - **Memory Details**: - Standard RSS (Resident Set Size) - PSS/USS via `/proc/PID/smaps` (requires root or `CAP_SYS_PTRACE`) - **Real PSS/USS**: Excludes file-backed mappings (accurate for apps that map large files) - **Systemd Services**: State, CPU, memory via cgroups (auto-discovery with filter) - **Docker Containers**: CPU, memory, network, disk I/O with optional rate metrics (KiB/s) - **Battery**: Level, status, voltage, current, health (auto-discovery supported) - **AC Power**: External power supply presence (`online`/`offline`, with auto-discovery) - **Network Interfaces**: Bytes, packets, errors, drops, rate, isup, speed, mtu, duplex, optional Wi-Fi RSSI (dBm) (auto-discovery supported) - **Fan (RPM)**: hwmon fan*_input from sysfs (auto-discovery supported) - **Custom Sensors**: Run shell commands or scripts - **Binary Sensors**: ON/OFF states from command execution (e.g., ping checks) - **GPU**: Basic metrics via sysfs (frequency, temperature) - minimal implementation ### MQTT Integration - **JSON Payloads**: Single topic per source with all metrics in JSON format - **Retain Modes**: `on` (retain all) or `off` (no retention) - **Availability**: Dual availability system (global app status + local source state) - **Last Will and Testament**: Automatic offline notification on disconnect ### Home Assistant Integration - **MQTT Discovery**: Automatic sensor and device registration - **Device Templates**: Define reusable device configurations with custom grouping - **Flexible Device Assignment**: Use `system`, `auto`, `none`, or templates for each sensor - **Value Templates**: Extract metrics from JSON payloads - **Stale Sensor Cleanup**: Removed sensors are automatically cleaned from Home Assistant ### Auto-Discovery - **Temperature Sensors**: Automatic detection with filter/exclude patterns - **Disk Partitions**: Auto-discovery of mounted block devices - **Batteries**: Auto-discovery of all power supplies - **AC Power Supplies**: Auto-discovery of non-battery power sources under `/sys/class/power_supply` - **Network Interfaces**: Auto-discovery of interfaces (filter/exclude by name, e.g. eth*, wlan*) - **Docker Containers**: Auto-discovery with name/image/label filters - **Systemd Services**: Auto-discovery with required filter (safety) - **Processes**: Auto-discovery with required filter (safety) - **Dynamic Refresh**: Periodic check for new/removed sources (configurable interval) ## Requirements - Python 3.11+ - Linux with `/proc` and `/sys` filesystems - MQTT broker (Mosquitto, EMQX, etc.) - Home Assistant with MQTT integration (optional) - Elevated privileges for some metrics (see [Permissions](#permissions) below) ## Installation ### Quick install via pip ```bash pip install penguin-metrics ``` This installs the `penguin-metrics` command and the `penguin_metrics` Python package. After installation, create a configuration file: ```bash sudo mkdir -p /etc/penguin-metrics # Download the example configuration curl -o /etc/penguin-metrics/config.conf \ https://raw.githubusercontent.com/clusterm/penguin-metrics/master/config.example.conf # Edit to match your setup sudo nano /etc/penguin-metrics/config.conf ``` ### Install from GitHub (latest) ```bash pip install git+https://github.com/clusterm/penguin-metrics.git ``` ### Install from source (development) ```bash git clone https://github.com/clusterm/penguin-metrics.git cd penguin-metrics # Create virtual environment (recommended) python3 -m venv .venv source .venv/bin/activate # Install in editable mode pip install -e . # Or with dev dependencies (ruff, mypy, pytest) pip install -e ".[dev]" ``` ### Running After installation (any method above), the `penguin-metrics` command becomes available: ```bash # Run directly penguin-metrics /etc/penguin-metrics/config.conf # Or via Python module python -m penguin_metrics /etc/penguin-metrics/config.conf ``` ### As systemd service ```bash # Copy the provided service file sudo cp penguin-metrics.service /etc/systemd/system/ # If you installed via pip globally, adjust ExecStart in the service file: # ExecStart=/usr/local/bin/penguin-metrics -v /etc/penguin-metrics/config.conf # If you used a virtualenv: # ExecStart=/opt/penguin-metrics/.venv/bin/penguin-metrics -v /etc/penguin-metrics/config.conf # Create config directory and copy example sudo mkdir -p /etc/penguin-metrics sudo cp config.example.conf /etc/penguin-metrics/config.conf sudo nano /etc/penguin-metrics/config.conf # Enable and start sudo systemctl daemon-reload sudo systemctl enable penguin-metrics sudo systemctl start penguin-metrics # Check status sudo systemctl status penguin-metrics sudo journalctl -u penguin-metrics -f ``` ### With Docker ```bash # Build image docker build -t penguin-metrics . # Run with docker-compose docker-compose up -d # Or run directly docker run -d \ --name penguin-metrics \ --privileged \ --pid=host \ -v /proc:/host/proc:ro \ -v /sys:/host/sys:ro \ -v /var/run/docker.sock:/var/run/docker.sock:ro \ -v ./config.conf:/etc/penguin-metrics/config.conf:ro \ --network=host \ penguin-metrics ``` ## Usage ### Command Line ```bash # Validate configuration penguin-metrics --validate /path/to/config.conf # Run with verbose logging penguin-metrics -v /path/to/config.conf # Run with debug logging penguin-metrics -d /path/to/config.conf # Show version penguin-metrics --version # Show help penguin-metrics --help ``` > **Note:** `penguin-metrics` and `python -m penguin_metrics` are equivalent. ### Configuration Validation ```bash $ penguin-metrics --validate config.conf Configuration summary: MQTT: 10.13.1.100:1833 Home Assistant Discovery: enabled System collectors: 1 Process monitors: 2 Service monitors: 1 Container monitors: 1 Battery monitors: 0 AC power monitors: 0 Custom sensors: 1 Configuration is valid! ``` ## Permissions Some metrics require elevated privileges to read `/proc/PID/io`, `/proc/PID/smaps`, and `/proc/PID/fd/` of other processes. Without sufficient permissions these metrics will silently remain empty (a one-time warning is logged at startup). | Metric | File | Required capability | |---|---|---| | Disk I/O (`disk`, `disk_rate`) for processes/services | `/proc/PID/io` | `CAP_DAC_READ_SEARCH` | | PSS/USS memory (`smaps`) | `/proc/PID/smaps` | `CAP_SYS_PTRACE` | | File descriptors (`fds`) | `/proc/PID/fd/` | `CAP_DAC_READ_SEARCH` | ### Option 1: Run as root The simplest approach. If running in Docker with `privileged: true` — no extra steps needed. ### Option 2: Grant capabilities to the Python binary This avoids running the entire process as root: ```bash # Find the Python binary PYTHON=$(readlink -f $(which python3)) # Grant required capabilities sudo setcap 'cap_dac_read_search,cap_sys_ptrace=ep' "$PYTHON" ``` To verify: ```bash getcap "$PYTHON" # /usr/bin/python3.13 cap_dac_read_search,cap_sys_ptrace=ep ``` To remove: ```bash sudo setcap -r "$PYTHON" ``` > **Note:** `setcap` is applied to the actual binary, not a symlink. Use `readlink -f` to resolve symlinks. After a Python upgrade the capability must be re-applied. ### Option 3: Grant capabilities to the systemd service If running as a systemd service under a dedicated user: ```ini [Service] User=penguin-metrics AmbientCapabilities=CAP_DAC_READ_SEARCH CAP_SYS_PTRACE ``` ### Option 4: Docker with specific capabilities Instead of `privileged: true`, grant only what is needed: ```yaml cap_add: - DAC_READ_SEARCH - SYS_PTRACE ``` ## Configuration Penguin Metrics uses an nginx-like configuration syntax with blocks and directives. ### Basic Structure ```nginx # Comments start with # block_type "optional_name" { directive value; directive value1 value2; nested_block { directive value; } } # Include other config files include "/etc/penguin-metrics/conf.d/*.conf"; ``` ### Common Directives All collector blocks support these directives: | Directive | Default | Description | |-----------|---------|-------------| | `display_name` | *(block name)* | Human-readable name for Home Assistant sensors. Does not affect topics or IDs | | `device` | *(varies)* | Device grouping: `system`, `auto`, `none`, or a template name | | `update_interval` | *(from defaults)* | Override the collection interval for this block | **Example** — `display_name` overrides only the display name in HA: ```nginx disk "nvme" { match name "nvme0n1p1"; display_name "NVME"; # → HA sensors: "Disk NVME Total", "Disk NVME Free", etc. } ``` Without `display_name`, sensor names use the block name: "Disk nvme Total", "Disk nvme Free", etc. ### MQTT Configuration ```nginx mqtt { host "localhost"; # MQTT broker address port 1883; # MQTT broker port username "user"; # Optional: username password "pass"; # Optional: password client_id "penguin"; # Optional: client ID (auto-generated if not set) topic_prefix "penguin_metrics"; # Base topic for all messages qos 1; # QoS level (0, 1, 2) retain on; # Retain messages keepalive 60; # Keepalive interval in seconds } ``` **Default values:** | Directive | Default | Description | |-----------|---------|-------------| | `host` | `"localhost"` | MQTT broker address | | `port` | `1883` (or `8883` if TLS on) | MQTT broker port | | `username` | *(none)* | Authentication username | | `password` | *(none)* | Authentication password | | `client_id` | *(auto-generated)* | MQTT client identifier | | `topic_prefix` | `"penguin_metrics"` | Base topic for all messages | | `qos` | `1` | Quality of Service (0, 1, 2) | | `retain` | `on` | Retain mode: `on` (retain all) or `off` (no retention) | | `keepalive` | `60` | Keepalive interval (seconds) | | `tls` | `off` | Enable TLS/SSL (use port 8883) | | `tls_insecure` | `off` | Skip server certificate verification (dev only) | | `cafile` | *(none)* | Path to CA certificate file | | `capath` | *(none)* | Path to CA certificates directory | | `certfile` | *(none)* | Path to client certificate file | | `keyfile` | *(none)* | Path to client private key file | **TLS/SSL:** When `tls on;` is set, the connection uses TLS. Specify `cafile` (or `capath`) to verify the broker certificate. For client certificate authentication, set `certfile` and `keyfile`. Use port 8883 (default when TLS is on). Set `tls_insecure on;` only for development to skip certificate verification. ```nginx # Example: MQTT over TLS mqtt { host "broker.example.com"; port 8883; tls on; cafile "/etc/ssl/certs/ca-certificates.crt"; # Optional: client certificate # certfile "/etc/penguin-metrics/client.crt"; # keyfile "/etc/penguin-metrics/client.key"; username "user"; password "pass"; topic_prefix "penguin_metrics"; } ``` **Retain modes:** | Mode | Description | |------|-------------| | `off` | Don't retain any messages | | `on` | Retain all messages (default) | When the service disconnects, `{"state": "offline"}` is sent to all JSON topics (if retain is enabled). ### Home Assistant Integration ```nginx homeassistant { discovery on; # Enable MQTT Discovery discovery_prefix "homeassistant"; # Discovery topic prefix } ``` **Default values:** | Directive | Default | Description | |-----------|---------|-------------| | `discovery` | `on` | Enable MQTT Discovery | | `discovery_prefix` | `"homeassistant"` | Discovery topic prefix | | `state_file` | `"/var/lib/penguin-metrics/registered_sensors.json"` | State file for stale sensor cleanup | ### Device Templates Define reusable device templates for grouping sensors in Home Assistant: ```nginx device "ups_batteries" { name "UPS Batteries"; manufacturer "APC"; model "Smart-UPS"; } ``` Reference templates in sensor configurations using `device "template_name";`: ```nginx battery "ups_battery_1" { device "ups_batteries"; # Use the template above # ... other settings } custom "room_temp" { device "room_sensors"; # Reference another template # ... other settings } ``` **Reserved device values:** | Value | Description | |-------|-------------| | `device system;` | Group with the system device (default for temperature, GPU, disks, battery) | | `device auto;` | Create a dedicated device (default for services, containers, processes, custom) | | `device none;` | No device — create "orphan" entities without device association | ### Home Assistant Sensor Overrides Override any Home Assistant discovery fields for sensors using the `homeassistant {}` block: ```nginx process "nginx" { homeassistant { name "Web Server"; # Override display name icon "mdi:web"; # Change icon unit_of_measurement "%"; # Override unit device_class "power"; # Change device class state_class "measurement"; # Change state class entity_category "diagnostic"; # Set entity category enabled_by_default false; # Disable by default # Any other HA discovery fields can be added here } match name "nginx"; cpu on; memory on; } ``` **Available fields in `homeassistant {}` block:** | Field | Description | |-------|-------------| | `name` | Display name in Home Assistant (default: auto-generated) | | `icon` | MDI icon name (e.g., `"mdi:thermometer"`) | | `unit_of_measurement` | Unit of measurement | | `device_class` | HA device class (e.g., `temperature`, `power`, `energy`) | | `state_class` | HA state class (e.g., `measurement`, `total`, `total_increasing`) | | `entity_category` | Entity category (`diagnostic`, `config`) | | `enabled_by_default` | Enable sensor by default (`true`/`false`, default: `true`) | | *Any other field* | Additional fields are passed directly to HA discovery payload | **Note:** The `homeassistant {}` block applies to all sensors created by the collector. For collectors that create multiple sensors (e.g., process, container), the overrides are applied to all of them. ### Default Settings ```nginx defaults { update_interval 10s; # Collection interval (supports: ms, s, m, h, d) smaps off; # PSS/USS memory metrics (requires root) } ``` **Default values:** | Directive | Default | Description | |-----------|---------|-------------| | `update_interval` | `10s` | Collection interval | | `smaps` | `off` | PSS/USS memory (requires root) | **Duration formats:** `ms` (milliseconds), `s` (seconds), `m` (minutes), `h` (hours), `d` (days) #### Per-Source-Type Defaults You can define default settings for each source type within the `defaults` block: ```nginx defaults { update_interval 10s; smaps off; # All processes will have these settings by default process { cpu on; memory on; smaps on; # Enable smaps for all processes disk on; # Read/write totals (bytes) fds off; threads off; } # All containers will have these settings by default container { cpu on; memory on; network on; # Enable network for all containers disk off; state on; health on; } # All services will have these settings by default service { cpu on; memory on; state on; restart_count on; } # All batteries will have these settings by default battery { level on; voltage on; power on; health on; } # All custom sensors will have these settings by default custom { type number; timeout 10s; } } ``` Individual source blocks can still override these defaults: ```nginx # Uses process defaults (smaps on from above) process "nginx" { match name "nginx"; } # Overrides process defaults (smaps off for this one) process "low-priority" { match name "background-task"; smaps off; } ``` ### Auto-Discovery Penguin Metrics can automatically discover sensors, batteries, AC power supplies, containers, services, processes, and disks. All auto-discovery settings are grouped inside the `auto_discovery { }` block: ```nginx auto_discovery { # Auto-discover temperature sensors temperatures { auto on; # source thermal; # "thermal" or "hwmon" (default: thermal) # update_interval 15s; # Optional override } # Auto-discover batteries batteries { auto on; # current off; # Optional per-metric override # temperature on; # Enable temperature metrics # update_interval 30s; } # Auto-discover running Docker containers (with filter) containers { auto on; filter "myapp-*"; # disk_rate on; # Override defaults.container values # update_interval 10s; } # Auto-discover systemd services (filter REQUIRED for safety) services { auto on; filter "docker*"; # REQUIRED - too many services otherwise # smaps on; } # Auto-discover processes (filter REQUIRED for safety) processes { auto on; filter "python*"; # REQUIRED - thousands of processes otherwise # smaps on; } # Auto-discover external power supplies (non-battery) ac_powers { auto on; # filter "axp*"; # exclude "usb*"; # update_interval 30s; } # Auto-discover network interfaces networks { auto on; # filter "eth*"; # Only Ethernet # exclude "lo"; # Exclude loopback # rate on; # Enable bytes rate (KiB/s) # update_interval 10s; } # Auto-discover disk partitions disks { auto on; filter "*"; # All partitions # exclude "loop*"; # update_interval 60s; } # Auto-discover fans (hwmon fan*_input) # fans { # auto on; # # filter "fan*"; # # update_interval 10s; # } } ``` **Multiple filters and excludes:** ```nginx auto_discovery { temperatures { auto on; source hwmon; # Use hwmon instead of thermal zones filter "nvme*"; # Include NVMe sensors filter "soc*"; # Include SoC sensors exclude "test*"; # Exclude test sensors } } ``` **Logic:** - If **any exclude** pattern matches → excluded - If filters defined and **any matches** → included - If no filters → include all (except excluded) - Auto blocks **inherit per-source defaults**, then apply any boolean flags and `update_interval` specified directly inside the auto block (e.g., `batteries { current off; temperature on; }`). ### Dynamic Auto-Refresh By default, auto-discovery runs only at startup. Enable `auto_refresh_interval` (top-level setting) to periodically check for new or removed sources: ```nginx # Check for new/removed sources every 60 seconds (0 = disabled) auto_refresh_interval 60s; ``` When enabled: - **New auto-discovered sources** (services, containers, processes, temperatures, batteries, disks, networks, fans) matching filters are automatically added - **Removed** auto-discovered sources are cleaned up from HA and JSON state - Home Assistant sensors are registered/unregistered dynamically - Manual configurations are never affected - Logs at `INFO` level only when sources are added/removed (not on every check) **Manual definitions override auto-discovered:** ```nginx # This overrides the auto-discovered "soc-thermal" temperature "soc-thermal" { match zone "soc-thermal"; update_interval 5s; } ``` **Stale sensor cleanup:** When a sensor disappears (e.g., NVMe removed), it will be automatically removed from Home Assistant on the next Penguin Metrics restart. State is stored in: - Primary: `/var/lib/penguin-metrics/registered_sensors.json` - Fallback: `~/.penguin-metrics/registered_sensors.json` Custom location: ```nginx homeassistant { state_file "/custom/path/sensors.json"; } ``` ### Logging Configuration ```nginx logging { level info; # Console log level colors on; # Colored console output (auto-detect TTY) # File logging file "/var/log/penguin-metrics/penguin-metrics.log"; file_level debug; # File log level file_max_size 10; # Max size in MB before rotation file_keep 5; # Number of rotated files to keep # Custom format (advanced) # format "%(asctime)s [%(levelname)s] %(name)s: %(message)s"; } ``` **Default values:** | Directive | Default | Description | |-----------|---------|-------------| | `level` | `"info"` | Console log level | | `colors` | `on` | Colored output (if TTY) | | `file` | *(none)* | Log file path | | `file_level` | `"debug"` | File log level | | `file_max_size` | `10` | Max file size (MB) | | `file_keep` | `5` | Backup files to keep | | `format` | `"%(asctime)s [%(levelname)s] %(name)s: %(message)s"` | Python logging format string | **Log levels:** `debug`, `info`, `warning`, `error` **Command-line overrides:** ```bash # Verbose (INFO level) penguin-metrics -v config.conf # Debug (DEBUG level) penguin-metrics -d config.conf # Quiet (ERROR only) penguin-metrics -q config.conf # Custom log file penguin-metrics --log-file /tmp/pm.log config.conf # Disable colors penguin-metrics --no-color config.conf ``` ### System Metrics ```nginx system "My Server" { # The system name becomes the device name in Home Assistant # Optional: device "template_name"; to use a device template cpu on; # Total CPU usage cpu_per_core off; # Per-core CPU usage memory on; # Memory usage swap on; # Swap usage load on; # Load average (1, 5, 15 min) uptime on; # System uptime gpu off; # GPU metrics (if available) update_interval 5s; } ``` **Note:** Temperature sensors are configured separately via `auto_discovery { temperatures { ... } }` (see Auto-Discovery section). **Default values:** | Directive | Default | Description | |-----------|---------|-------------| | `cpu` | `on` | Total CPU usage | | `cpu_per_core` | `off` | Per-core CPU usage | | `memory` | `on` | Memory usage | | `swap` | `on` | Swap usage | | `load` | `on` | Load average | | `uptime` | `on` | System uptime | | `gpu` | `off` | GPU metrics | | `disk_io` | `on` | Disk read/write totals (bytes) | | `disk_io_rate` | `off` | Disk read/write rate (KiB/s) | | `cpu_freq` | `on` | CPU frequency current/min/max (MHz; N/A on some ARM/virtual) | | `process_count` | `on` | Total and running process count | | `boot_time` | `on` | Boot time (timestamp for HA) | | `update_interval` | *(from defaults)* | Override default interval | ### Process Monitoring ```nginx # Match by exact process name process "docker" { match name "dockerd"; cpu on; memory on; smaps on; # PSS/USS + Real PSS/USS (requires root) disk on; # Read/write totals (bytes) fds on; # Open file descriptors threads on; # Thread count } # Match by regex pattern process "nginx-workers" { match pattern "nginx: worker.*"; aggregate on; # Sum metrics from all matches cpu on; memory on; } # Match by PID file process "my-app" { match pidfile "/var/run/my-app.pid"; cpu on; memory on; } # Match by command line substring process "python-script" { match cmdline "/opt/scripts/main.py"; cpu on; memory on; } ``` **Default values:** | Directive | Default | Description | |-----------|---------|-------------| | `cpu` | `on` | CPU usage (normalized to 0-100%) | | `memory` | `on` | Memory (RSS) | | `smaps` | *(from defaults)* | PSS/USS + Real PSS/USS memory | | `disk` | `off` | Read/write totals (bytes) | | `disk_rate` | `off` | Read/write rate (KiB/s) | | `fds` | `off` | Open file descriptors (enable for open files monitoring) | | `threads` | `off` | Thread count (enable for thread monitoring) | | `aggregate` | `off` | Sum metrics from all matches | | `update_interval` | *(from defaults)* | Override default interval | **Memory metrics when `smaps on`:** - `memory_rss_mb`: Standard RSS (Resident Set Size) - `memory_pss_mb`: Proportional Set Size (includes file-backed) - `memory_uss_mb`: Unique Set Size (includes file-backed) - `memory_real_pss_mb`: Real PSS (excludes file-backed mappings) - `memory_real_uss_mb`: Real USS (excludes file-backed mappings) **Match types:** | Type | Example | Description | |------|---------|-------------| | `name` | `match name "nginx";` | Exact process name (comm) | | `pattern` | `match pattern "nginx:.*";` | Regex on command line | | `pid` | `match pid 1234;` | Exact PID | | `pidfile` | `match pidfile "/var/run/app.pid";` | Read PID from file | | `cmdline` | `match cmdline "/usr/bin/app";` | Substring in command line | **State field** (always included): `running` (process(es) found and metrics collected), `not_found` (no matching processes), `error` (e.g. access denied). ### Systemd Service Monitoring ```nginx service "docker" { match unit "docker.service"; cpu on; # CPU time from cgroup memory on; # Memory Cgroup (includes cache, use smaps for accurate) smaps on; # PSS/USS aggregated state on; # active/inactive/failed restart_count on; # Number of restarts } # Match by pattern service "nginx" { match pattern "nginx*.service"; cpu on; memory on; state on; } ``` **Default values:** | Directive | Default | Description | |-----------|---------|-------------| | `cpu` | `on` | CPU usage (normalized to 0-100%) | | `memory` | `on` | Memory Cgroup (includes cache) | | `smaps` | *(from defaults)* | PSS/USS + Real PSS/USS aggregated | | `state` | `on` | Service state (only 'active' collects metrics) | | `restart_count` | `off` | Number of restarts | | `disk` | `off` | Read/write totals (bytes) | | `disk_rate` | `off` | Read/write rate (KiB/s) | | `update_interval` | *(from defaults)* | Override default interval | **Note:** Metrics are only collected when service state is `active`. States like `activating` or `reloading` don't collect cgroup metrics. **State field** (when `state on`): systemd ActiveState — `active`, `inactive`, `failed`, `activating`, `deactivating`, `reloading`; or `not_found` (unit missing), `unknown` (property unreadable). **Match types:** | Type | Example | Description | |------|---------|-------------| | `unit` | `match unit "docker.service";` | Exact unit name | | `pattern` | `match pattern "nginx*.service";` | Glob pattern | ### Docker Container Monitoring ```nginx container "homeassistant" { match name "homeassistant"; cpu on; # CPU usage % memory on; # Memory usage network on; # Network RX/TX disk on; # Block I/O state on; # running/exited/paused/etc health on; # Healthcheck status uptime on; # Container uptime } # Match by image container "postgres" { match image "postgres:"; cpu on; memory on; state on; } # Match by label container "monitored" { match label "metrics.enabled=true"; cpu on; memory on; } ``` **Default values:** | Directive | Default | Description | |-----------|---------|-------------| | `cpu` | `on` | CPU usage % (normalized to 0-100%) | | `memory` | `on` | Memory usage | | `network` | `off` | Network RX/TX bytes | | `network_rate` | `off` | Network RX/TX rate (KiB/s) | | `disk` | `off` | Block I/O | | `disk_rate` | `off` | Block I/O rate (KiB/s) | | `state` | `on` | Container state | | `health` | `off` | Healthcheck status | | `uptime` | `off` | Container uptime | | `update_interval` | *(from defaults)* | Override default interval | **Match types:** | Type | Example | Description | |------|---------|-------------| | `name` | `match name "nginx";` | Exact container name | | `pattern` | `match pattern "web-.*";` | Regex on name | | `image` | `match image "postgres:";` | Image name (substring) | | `label` | `match label "app=web";` | Container label | **State field** (when `state on`): Docker container state — `running`, `exited`, `paused`, `restarting`, `dead`, `created`, `removing`; or `not_found` (container missing), `unknown` (API did not return state). ### Battery Monitoring **Metrics (published as JSON fields):** - `state` - charging/discharging/full/not charging/not_found - `level` - battery level (%) - `voltage` - current voltage (V) - `current` - current (A, sign preserved) - `power` - power (W, sign preserved) - `health` - battery health - `cycles` - charge cycle count - `temperature` - battery temperature (°C) - `time_to_empty` - minutes remaining - `time_to_full` - minutes to full charge - `energy_now`, `energy_full`, `energy_full_design` - energy (Wh) - `present` - presence flag (0/1) - `technology` - chemistry (e.g., Li-ion) - `voltage_max`, `voltage_min` - current voltage limits (V) - `voltage_max_design`, `voltage_min_design` - design voltage limits (V) - `constant_charge_current`, `constant_charge_current_max` - charge currents (A) - `charge_full_design` - design full charge (mAh) ```nginx battery "main" { # Match criteria (exactly one): match name "BAT0"; # Battery name # match path "/sys/class/power_supply/BAT0"; # Or by sysfs path level on; # Battery level (%) voltage on; # Current voltage current on; # Current amperage (sign preserved: +charge / -discharge) power on; # Power (sign preserved) health on; # Battery health energy_now on; # Current energy (Wh) energy_full on; # Full energy (Wh) energy_full_design on; # Design full energy (Wh) cycles on; # Charge cycles temperature on; # Battery temperature time_to_empty on; # Estimated time remaining time_to_full on; # Time to full charge present on; # Presence flag (0/1) technology on; # Chemistry (e.g., Li-ion) voltage_max on; # Current max voltage (V) voltage_min on; # Current min voltage (V) voltage_max_design on; # Design max voltage (V) voltage_min_design on; # Design min voltage (V) constant_charge_current on; # Target charge current (A) constant_charge_current_max on; # Max charge current (A) charge_full_design on; # Design full charge (mAh) update_interval 30s; } ``` **Default values:** | Directive | Default | Description | |-----------|---------|-------------| | `match name` | *(required)* | Battery name (BAT0, etc.) | | `match path` | *(alternative)* | Full path to battery in sysfs | | `level` | `on` | Battery level (%) | | `voltage` | `on` | Current voltage | | `current` | `on` | Current amperage (sign preserved) | | `power` | `on` | Power (sign preserved) | | `health` | `on` | Battery health | | `energy_now` | `on` | Current energy (Wh) | | `energy_full` | `on` | Full energy (Wh) | | `energy_full_design` | `on` | Design full energy (Wh) | | `cycles` | `off` | Charge cycle count | | `temperature` | `off` | Battery temperature | | `time_to_empty` | `off` | Time remaining | | `time_to_full` | `off` | Time to full charge | | `present` | `off` | Presence flag | | `technology` | `off` | Battery chemistry | | `voltage_max` | `off` | Current max voltage | | `voltage_min` | `off` | Current min voltage | | `voltage_max_design` | `off` | Design max voltage | | `voltage_min_design` | `off` | Design min voltage | | `constant_charge_current` | `off` | Target charge current | | `constant_charge_current_max` | `off` | Max charge current | | `charge_full_design` | `off` | Design full charge (mAh) | | `update_interval` | *(from defaults)* | Override default interval | Status (`state`) is always collected and published (required for availability/HA) and is not configurable. ### AC Power Monitoring Monitors external power supply (mains) presence from `/sys/class/power_supply/<device>/online`. **Metrics (published as JSON fields):** - `state` - online/not_found (source availability: "online" if data read successfully, "not_found" if source unavailable) - `online` - boolean: `true` if external power is present, `false` otherwise ```nginx ac_power "main" { # Match criteria (exactly one): match name "axp22x-ac"; # Sysfs device name # match path "/sys/class/power_supply/axp22x-ac"; # Or by full path device system; # Group with system device (default) # update_interval 30s; # homeassistant { name "AC Power"; icon "mdi:power-plug"; } } ``` **Default values:** | Directive | Default | Description | |-----------|---------|-------------| | `match name` | *(required)* | Sysfs device name (e.g. axp22x-ac) | | `match path` | *(alternative)* | Full path to power_supply directory | | `device` | `system` | Group with system device (via parent device) | | `update_interval` | *(from defaults)* | Override default interval | **Note:** AC power sensors publish JSON with `online` (boolean) and `state` fields. Exposed to Home Assistant as a `binary_sensor` with `ON`/`OFF` derived from `online`. ### Disk Space Monitors disk space usage from mounted partitions via `psutil.disk_usage()`. **Metrics (published as JSON fields):** - `total` - Total size (GiB) - `used` - Used space (GiB) - `free` - Free space (GiB) - `percent` - Usage percentage (%) - `state` - online/not_found (source availability) ```nginx disk "root" { # Match criteria (exactly one): match name "sda1"; # Device name (from /dev/) # match mountpoint "/"; # Or by mountpoint # match uuid "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx"; # Or by UUID # Display name for Home Assistant (optional, defaults to block name) # display_name "Root Disk"; device system; # Group with system device (default) total on; # Total size (GiB) used on; # Used space (GiB) free on; # Free space (GiB) percent on; # Usage percentage # update_interval 60s; } ``` **Default values:** | Directive | Default | Description | |-----------|---------|-------------| | `match name` | *(required)* | Device name (e.g. sda1, nvme0n1p1) | | `match mountpoint` | *(alternative)* | Mountpoint (e.g. /, /home) | | `match uuid` | *(alternative)* | UUID from `/dev/disk/by-uuid/` | | `total` | `on` | Total size (GiB) | | `used` | `on` | Used space (GiB) | | `free` | `on` | Free space (GiB) | | `percent` | `on` | Usage percentage | | `device` | `system` | Group with system device | | `update_interval` | *(from defaults)* | Override default interval | **Match types:** | Type | Example | Description | |------|---------|-------------| | `name` | `match name "nvme0n1p1";` | Device name (from `/dev/`) | | `mountpoint` | `match mountpoint "/home";` | Mountpoint path | | `uuid` | `match uuid "a1b2c3d4-...";` | UUID (stable across reboots) | ### Network Interfaces Monitors network interfaces via `psutil.net_io_counters(pernic=True)` and `psutil.net_if_stats()`. **Metrics (published as JSON fields):** - `bytes_sent`, `bytes_recv` - Total bytes - `packets_sent`, `packets_recv` - Packet counts - `errin`, `errout` - Error counts - `dropin`, `dropout` - Dropped packet counts - `bytes_sent_rate`, `bytes_recv_rate` - Rate (KiB/s) when `rate on` - `packets_sent_rate`, `packets_recv_rate` - Packet rate (p/s) when `packets_rate on` - `isup` - Interface up/down (boolean, binary_sensor in HA) - `speed` - Link speed (Mbps) - `mtu` - MTU - `duplex` - full/half - `state` - online/not_found (source availability) ```nginx network "eth0" { match name "eth0"; # Interface name (required) device system; # Group with system device (default) bytes on; # bytes_sent, bytes_recv (bytes) packets off; # packets_sent, packets_recv errors off; # errin, errout drops off; # dropin, dropout rate off; # bytes_sent_rate, bytes_recv_rate (KiB/s) packets_rate off; # packets_sent_rate, packets_recv_rate (p/s) isup on; # Interface up/down (binary_sensor) speed off; # Speed (Mbps) mtu off; duplex off; rssi off; # Wi-Fi signal (dBm) for wireless interfaces # update_interval 10s; } ``` **Default values (defaults.network):** `bytes` on, `packets`/`errors`/`drops`/`rate`/`packets_rate` off, `isup` on, `speed`/`mtu`/`duplex` off, `rssi` off. Optional `rssi on`: Wi-Fi signal strength (dBm) for wireless interfaces (uses `iw` or `iwconfig`). ### Fan (RPM) Fan speed from hwmon sysfs (`/sys/class/hwmon/hwmon*/fan*_input`). Manual config or auto-discovery via `auto_discovery { fans { auto on; } }`. ```nginx fan "cpu_fan" { match hwmon "hwmon0"; # Hwmon directory name (required) device system; } ``` Metrics: `fan1_rpm`, `fan2_rpm`, or `rpm` (single fan). Unit: RPM. ### Custom Sensors The block name (e.g., `"room_temp"`) is the sensor ID, used for MQTT topics. Use the `homeassistant {}` block to override any Home Assistant discovery fields. ```nginx # Read from command output # MQTT topic: {prefix}/custom/room_temp custom "room_temp" { command "cat /sys/bus/w1/devices/28-*/temperature"; type number; # number, string, json scale 0.001; # Multiply result # Home Assistant sensor overrides homeassistant { name "Room Temperature"; # Display name in HA (default: use ID) icon "mdi:thermometer"; unit_of_measurement "°C"; device_class temperature; state_class measurement; # Any other HA discovery fields can be added here } update_interval 30s; timeout 5s; } # Run script custom "disk_health" { script "/opt/scripts/check_smart.sh"; type string; update_interval 1h; } # Get external IP custom "wan_ip" { homea
text/markdown
Alexey Cluster, Opus 4.5
null
null
null
GPLv3
mqtt, home-assistant, telemetry, monitoring, linux, metrics
[ "Development Status :: 4 - Beta", "Environment :: No Input/Output (Daemon)", "Intended Audience :: System Administrators", "License :: OSI Approved :: GNU General Public License v3 (GPLv3)", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3...
[]
null
null
>=3.11
[]
[]
[]
[ "aiomqtt>=2.0.0", "psutil>=5.9.0", "pytest>=7.0.0; extra == \"dev\"", "pytest-asyncio>=0.21.0; extra == \"dev\"", "pytest-cov>=4.0.0; extra == \"dev\"", "ruff>=0.1.0; extra == \"dev\"", "mypy>=1.0.0; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://github.com/clusterm/penguin-metrics", "Documentation, https://github.com/clusterm/penguin-metrics#readme", "Repository, https://github.com/clusterm/penguin-metrics", "Issues, https://github.com/clusterm/penguin-metrics/issues" ]
twine/6.2.0 CPython/3.13.5
2026-02-19T12:00:09.885582
penguin_metrics-0.0.2.tar.gz
142,362
7e/aa/ae8b206a7e276702a039e12c958e9267e30ead89ed8f3199a0ccf44c4c43/penguin_metrics-0.0.2.tar.gz
source
sdist
null
false
abd5a36dc1246d9d2593549296ee5436
28864f94d84e52911143c67d856445de042c176d94796bba8a61544d0a4ae5d5
7eaaae8b206a7e276702a039e12c958e9267e30ead89ed8f3199a0ccf44c4c43
null
[ "LICENSE" ]
239
2.1
py-import-cycles
0.5.2
Detect import cycles in Python projects
py-import-cycles ================ Detect import cycles in Python projects. It: * walks over given packages, * collects (file-based) Python modules, * extracts import statements via `ast` and * computes cycles. It is conceived for having an indication whether Python packages may have structural weak points. `py-import-cycles` does not take any Python module finder or loader mechanisms into account. Installation ------------ The py-import-cycles package is available on PyPI: `python3 -m pip install --user py-import-cycles` Usage ----- * `python3 -m py_import_cycles --version` * `python3 -m py_import_cycles --help` * `python3 -m py_import_cycles --packages /path/to/project/package`
text/markdown
Simon Jess
simon86betz@yahoo.de
null
null
MIT
null
[ "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.12" ]
[]
https://github.com/si-23/py-import-cycles
null
null
[]
[]
[]
[]
[]
[]
[]
[]
twine/5.0.0 CPython/3.12.6
2026-02-19T11:59:57.385437
py_import_cycles-0.5.2.tar.gz
13,099
8e/74/3c84a5fa699e0ac1b8a5c4e1ca36dd3afc62aa7ff2e3b25ac6dc0eaf7bf5/py_import_cycles-0.5.2.tar.gz
source
sdist
null
false
bd14af570684ea37ef5bc6789aa97112
f1ae388c12e8cfbb0586405110574aff85ab9d570111319cc31ef4f57659264c
8e743c84a5fa699e0ac1b8a5c4e1ca36dd3afc62aa7ff2e3b25ac6dc0eaf7bf5
null
[]
4,369
2.4
python-log-viewer
1.0.3
A beautiful, real-time log viewer with a web UI. Works with Django, Flask, and FastAPI.
# python-log-viewer [![PyPI Downloads](https://static.pepy.tech/personalized-badge/python-log-viewer?period=total&units=INTERNATIONAL_SYSTEM&left_color=BLACK&right_color=GREEN&left_text=Downloads)](https://pepy.tech/projects/python-log-viewer) A beautiful, real-time log viewer with a dark-themed web UI. Browse, search, filter, clear, and delete log files — all from your browser. Integrates seamlessly with **Django**, **Flask**, and **FastAPI**. ## Preview ![Log Viewer Preview](https://raw.githubusercontent.com/imsujan276/python-log-viewer/main/screenshots/preview.png) ## Support If you like the package and find it helpful, you can [Buy Me MO:MO](https://buymemomo.com/sujang). --- ## Features - 📁 **File browser** — sidebar with folder tree, file sizes - 🔍 **Search & filter** — full-text search, log-level filtering (DEBUG / INFO / WARNING / ERROR) - 🎨 **Colour-coded** — log levels highlighted with subtle background colours - 🔄 **Auto-refresh** — configurable live-tail (5s, 10s, 30s, 1m, or manual) - 📜 **Line limits** — last 500 / 1000 / 2500 / 5000 / all entries - 🗑️ **File actions** — clear (truncate) or delete log files with confirmation modals - 🔒 **Basic Auth** — optional HTTP Basic Authentication - 📱 **Responsive** — works on mobile with a slide-out sidebar --- ## Installation ```bash pip install python-log-viewer ``` ### Framework extras ```bash pip install python-log-viewer[django] # Django integration pip install python-log-viewer[flask] # Flask integration pip install python-log-viewer[fastapi] # FastAPI integration pip install python-log-viewer[all] # All frameworks ``` --- ## Django Integration ### 1. Install ```bash pip install python-log-viewer[django] ``` ### 2. Add to `INSTALLED_APPS` ```python INSTALLED_APPS = [ # ... "python_log_viewer.contrib.django", ] ``` ### 3. Include URLs ```python # urls.py from django.urls import path, include urlpatterns = [ # ... path("logs/", include("python_log_viewer.contrib.django.urls")), ] ``` ### 4. Configure (optional) Add any of these to your `settings.py`: ```python # Path to your log directory (default: BASE_DIR / "logs") LOG_VIEWER_DIR = BASE_DIR / "logs" # UI defaults LOG_VIEWER_AUTO_REFRESH = True # enable auto-refresh LOG_VIEWER_REFRESH_TIMER = 5000 # refresh interval in ms LOG_VIEWER_AUTO_SCROLL = True # auto-scroll to bottom LOG_VIEWER_COLORIZE = True # colour-coded log levels # Authentication (optional — leave unset to disable) LOG_VIEWER_USERNAME = "admin" LOG_VIEWER_PASSWORD = "secret" # Allow logged-in Django superusers to bypass Basic Auth (default: True) LOG_VIEWER_SUPERUSER_ACCESS = True ``` Then visit `http://localhost:8000/logs/` in your browser. --- ## Flask Integration ### 1. Install ```bash pip install python-log-viewer[flask] ``` ### 2. Register the blueprint ```python from flask import Flask from python_log_viewer.contrib.flask import create_log_viewer_blueprint app = Flask(__name__) app.register_blueprint( create_log_viewer_blueprint( log_dir="./logs", url_prefix="/logs", username="admin", # optional password="secret", # optional ) ) if __name__ == "__main__": app.run(debug=True) ``` Then visit `http://localhost:5000/logs/` in your browser. **Blueprint parameters:** | Parameter | Default | Description | |-----------|---------|-------------| | `log_dir` | `"./logs"` | Path to log directory | | `url_prefix` | `"/logs"` | URL prefix | | `username` | `None` | Basic-Auth username | | `password` | `None` | Basic-Auth password | | `auto_refresh` | `True` | Enable auto-refresh | | `refresh_timer` | `5000` | Refresh interval (ms) | | `auto_scroll` | `True` | Auto-scroll to bottom | | `colorize` | `True` | Colour-coded levels | --- ## FastAPI Integration ### 1. Install ```bash pip install python-log-viewer[fastapi] ``` ### 2. Include the router ```python from fastapi import FastAPI from python_log_viewer.contrib.fastapi import create_log_viewer_router app = FastAPI() app.include_router( create_log_viewer_router( log_dir="./logs", prefix="/logs", username="admin", # optional password="secret", # optional ) ) ``` Then visit `http://localhost:8000/logs/` in your browser. **Router parameters:** | Parameter | Default | Description | |-----------|---------|-------------| | `log_dir` | `"./logs"` | Path to log directory | | `prefix` | `"/logs"` | URL prefix | | `username` | `None` | Basic-Auth username | | `password` | `None` | Basic-Auth password | | `auto_refresh` | `True` | Enable auto-refresh | | `refresh_timer` | `5000` | Refresh interval (ms) | | `auto_scroll` | `True` | Auto-scroll to bottom | | `colorize` | `True` | Colour-coded levels | --- ## Using the Core API Directly The core classes have **zero dependencies** and can be used in any Python application: ```python from python_log_viewer.core import LogDirectory, LogReader # Point to your log directory log_dir = LogDirectory("/var/log/myapp") # List all files for f in log_dir.list_files(): print(f"{f.name} {f.size} bytes modified={f.modified}") # Read and filter log entries reader = LogReader(log_dir) result = reader.read( file="app.log", lines=100, level="ERROR", search="database", ) print(f"Total matching entries: {result['total']}") for line in result["lines"]: print(line) # File operations log_dir.clear_file("app.log") # truncate to 0 bytes log_dir.delete_file("old.log") # permanently remove ``` --- ## Environment Variables Configuration can be set via environment variables (useful for Docker / CI): | Variable | Description | |----------|-------------| | `LOG_VIEWER_USERNAME` | Basic-Auth username | | `LOG_VIEWER_PASSWORD` | Basic-Auth password | --- ## Development ```bash # Clone git clone https://github.com/imsujan276/python-log-viewer.git cd python-log-viewer # Install in editable mode pip install -e ".[all]" ``` --- ## License MIT
text/markdown
null
Sujan Gainju <sujangainju01@gmail.com>
null
null
MIT
django, fastapi, flask, log, logging, monitoring, viewer, web-ui
[ "Development Status :: 4 - Beta", "Framework :: Django", "Framework :: FastAPI", "Framework :: Flask", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.9", ...
[]
null
null
>=3.9
[]
[]
[]
[ "django>=3.2; extra == \"all\"", "fastapi>=0.68; extra == \"all\"", "flask>=2.0; extra == \"all\"", "uvicorn>=0.15; extra == \"all\"", "django>=3.2; extra == \"django\"", "fastapi>=0.68; extra == \"fastapi\"", "uvicorn>=0.15; extra == \"fastapi\"", "flask>=2.0; extra == \"flask\"" ]
[]
[]
[]
[ "Homepage, https://github.com/imsujan276/python-log-viewer", "Repository, https://github.com/imsujan276/python-log-viewer", "Issues, https://github.com/imsujan276/python-log-viewer/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-19T11:59:03.123476
python_log_viewer-1.0.3.tar.gz
16,435
5c/5d/4f90bc9d0898897de6d58531d1ecd317537227709bfed64441cc9e234170/python_log_viewer-1.0.3.tar.gz
source
sdist
null
false
dbb01ba92047bb47bae3f145e7afcd09
67efc02b45a9dbcab86a49c3b7afaa704011ff0f462df8975430a37cd1cd50c8
5c5d4f90bc9d0898897de6d58531d1ecd317537227709bfed64441cc9e234170
null
[]
239
2.1
spirack
0.2.10
Drivers for the QuTech SPI-rack
# SPI Rack The SPI Rack is a modular electronic instrumentation platform developed by QuTech. It has been developed to perform measurements on nanoelectronic devices, but is not limited to this. Design priority was the minimization of noise and interference signals on the wires connected to the measured device (sample). To learn more about the SPI Rack, use cases and the available modules, browse the [homepage](http://qtwork.tudelft.nl/~mtiggelman/). This repository contains the Python code to interface with the hardware. All the low level communication is handled by the classes and the user is presented with an easy interface to control the modules. Here is a simple example on how to use the D5a (16 channel 18-bit DAC module) to show how easy it is to get going: ```Python # Import parts of the SPI Rack library from spirack import SPI_rack, D5a_module # Instantiate the controller module spi = SPI_rack(port="COM4", baud=9600, timeout=1) # Unlock the controller for communication to happen spi.unlock() # Instantiate the D5a module using the controller module # and the correct module address D5a = D5a_module(spi, module=2) # Set the output of DAC 1 to the desired voltage D5a.set_voltage(0, voltage=2.1) ``` More examples can be found as Jupyter notebooks in [examples](https://github.com/mtiggelman/SPI-rack/tree/master/examples) or at the [website](http://qtwork.tudelft.nl/~mtiggelman/software/examples.html). ## Installation **Windows 7&8 users:** before connecting the SPI-rack for the first time, install the drivers located in `drivers.zip`. On 64-bit systems run `SPI-Rack_x64`, on 32-bit systems run `SPI-Rack_x86`. This is not necessary anymore for Windows 10 systems. For a basic install use: `pip install spirack`. For more details see the website [here](http://qtwork.tudelft.nl/~mtiggelman/software/setup.html). ## Qcodes Qcodes wrappers for certain modules are available from https://github.com/QCoDeS/Qcodes ## License See [License](https://github.com/mtiggelman/SPI-rack/blob/master/LICENSE).
text/markdown
Marijn Tiggelman
qutechdev@gmail.com
null
null
MIT
SPI, Qcodes, SPI-rack, QuTech, TU Delft, SPI
[ "Programming Language :: Python :: 3", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent" ]
[]
https://github.com/mtiggelman/SPI-rack
null
null
[]
[]
[]
[ "pyserial", "numpy" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.11.7
2026-02-19T11:58:48.467804
spirack-0.2.10.tar.gz
58,027
64/b9/fc1afd927ca80e770a024c24faa435b367d5e9a9021b048d8a11e9dee9cb/spirack-0.2.10.tar.gz
source
sdist
null
false
52218b3453af5e1c8a7760316d439a76
e9bfa50526026a45aaa212a8a0daaebca34d8332d89f909bb8eb6a391991025a
64b9fc1afd927ca80e770a024c24faa435b367d5e9a9021b048d8a11e9dee9cb
null
[]
1,427
2.4
ecquote
2.0.4
A package to estimate the cost and volume of an ECMWF dissemination feed
# ecquote A package to estimate the cost and volume of an ECMWF dissemination feed based on the rules described [there](https://www.ecmwf.int/en/forecasts/accessing-forecasts/payment-rules-and-options/tariffs). To use the tool, edit a file that contains the a series of [MARS requests](https://confluence.ecmwf.int/display/UDOC/MARS+command+and+request+syntax) as follows: class=od, stream=oper, expver=0001, levtype=pl, type=an, time=0000/1200, param=z/t/u/v, levelist=500/850, grid=0.1/0.1, area=40/0/30/50 type=fc, step=0/to/240/by/24 type=cf, stream=enfo type=pf, number=1/to/50 The verb of the request, such as `retrieve` or `dissemination`, is optional. The example above describes four requests. As for with the MARS language, attributes that are not specified are inherited from the previous request. Then simply run the `ecquote` tools: ecquote myrequests.req You then should get a result such as: Grand total: Yearly volume: 256.9 GiB Yearly fields: 3,346,320 Volume band: 1 GiB Volume cost: EUR 200 EPUs: 16,627 Discounted EPUs: 10,776 Information cost: EUR 5,388 (core: 5,388) To get a description of how the cost is determined, rerun the command with the option `--detailed`. # Conditions of use This tool is provided for evaluation and review purchases only. Only quotes submitted via the Product Requirements Catalogue will be reviewed/considered/approved: https://apps.ecmwf.int/shopping-cart/ # Licence Copyright 2024 European Centre for Medium-Range Weather Forecasts (ECMWF). Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at: http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.
text/markdown
European Centre for Medium-Range Weather Forecasts (ECMWF)
software.support@ecmwf.int
null
null
Apache License Version 2.0
tool
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Lan...
[]
https://github.com/ecmwf/ecquote
null
null
[]
[]
[]
[ "pyyaml", "tqdm", "xlsxwriter" ]
[]
[]
[]
[]
twine/6.2.0 CPython/3.14.3
2026-02-19T11:58:05.012881
ecquote-2.0.4.tar.gz
619,514
84/78/900a3d4a4a5ab26d47bfbe492b8ae84ff8c43277ebc8f6bf5b9c8b191b93/ecquote-2.0.4.tar.gz
source
sdist
null
false
4681bf245e9aca56a2c16f9db4d56983
3a53d7b87627eff04a522bcc48c660590df68a33e6000d2a9ed8764dd0150ab2
8478900a3d4a4a5ab26d47bfbe492b8ae84ff8c43277ebc8f6bf5b9c8b191b93
null
[ "LICENSE" ]
173
2.4
pmxt
2.10.0
Unified prediction market data API - The ccxt for prediction markets
# PMXT Python SDK A unified Python interface for interacting with multiple prediction market exchanges (Kalshi, Polymarket). > **Note**: This SDK requires the PMXT sidecar server to be running. See [Installation](#installation) below. ## Installation ```bash pip install pmxt ``` **Prerequisites**: The Python SDK requires the PMXT server, which is distributed via npm: ```bash npm install -g pmxtjs ``` That's it! The server will start automatically when you use the SDK. ## Quick Start ```python import pmxt # Initialize exchanges (server starts automatically!) poly = pmxt.Polymarket() kalshi = pmxt.Kalshi() # Search for markets markets = poly.fetch_markets(query="Trump") print(markets[0].title) # Get outcome details outcome = markets[0].outcomes[0] print(f"{outcome.label}: {outcome.price * 100:.1f}%") # Fetch historical data (use outcome.outcome_id!) candles = poly.fetch_ohlcv( outcome.outcome_id, resolution="1d", limit=30 ) # Get current order book order_book = poly.fetch_order_book(outcome.outcome_id) spread = order_book.asks[0].price - order_book.bids[0].price print(f"Spread: {spread * 100:.2f}%") ``` ### How It Works The Python SDK automatically manages the PMXT sidecar server: 1. **First API call**: Checks if server is running 2. **Auto-start**: Starts server if needed (takes ~1-2 seconds) 3. **Reuse**: Multiple Python processes share the same server 4. **Zero config**: Just import and use! ### Manual Server Control (Optional) If you prefer to manage the server yourself: ```python # Disable auto-start poly = pmxt.Polymarket(auto_start_server=False) # Or start the server manually in a separate terminal # $ pmxt-server ``` ## Authentication (for Trading) ### Polymarket Requires your **Polygon Private Key**: ```python import os import pmxt poly = pmxt.Polymarket( private_key=os.getenv("POLYMARKET_PRIVATE_KEY"), proxy_address=os.getenv("POLYMARKET_PROXY_ADDRESS"), # Optional # signature_type='gnosis-safe' (default) ) # Check balance balances = poly.fetch_balance() print(f"Available: ${balances[0].available}") # Place order (using outcome shorthand) markets = poly.fetch_markets(query="Trump") order = poly.create_order( outcome=markets[0].yes, side="buy", type="limit", amount=10, price=0.55 ) ``` ### Kalshi Requires **API Key** and **Private Key**: ```python import os import pmxt kalshi = pmxt.Kalshi( api_key=os.getenv("KALSHI_API_KEY"), private_key=os.getenv("KALSHI_PRIVATE_KEY") ) # Check positions positions = kalshi.fetch_positions() for pos in positions: print(f"{pos.outcome_label}: ${pos.unrealized_pnl:.2f}") ``` ### Limitless Requires **Private Key**: ```python import os import pmxt limitless = pmxt.Limitless( private_key=os.getenv("LIMITLESS_PRIVATE_KEY") ) # Check balance balances = limitless.fetch_balance() print(f"Available: ${balances[0].available}") ``` ## API Reference ### Market Data Methods - `fetch_markets(params?)` - Get active markets ```python # Fetch recent markets poly.fetch_markets(limit=20, sort='volume') # Search by text poly.fetch_markets(query='Fed rates', limit=10) # Fetch by slug/ticker poly.fetch_markets(slug='who-will-trump-nominate-as-fed-chair') ``` - `filter_markets(markets, query)` - Filter markets by keyword - `fetch_ohlcv(outcome_id, params)` - Get historical price candles - `fetch_order_book(outcome_id)` - Get current order book - `fetch_trades(outcome_id, params)` - Get trade history - `get_execution_price(order_book, side, amount)` - Get execution price - `get_execution_price_detailed(order_book, side, amount)` - Get detailed execution info ### Trading Methods (require authentication) - `create_order(params)` - Place a new order - `cancel_order(order_id)` - Cancel an open order - `fetch_order(order_id)` - Get order details - `fetch_open_orders(market_id?)` - Get all open orders ### Account Methods (require authentication) - `fetch_balance()` - Get account balance - `fetch_positions()` - Get current positions ## Data Models All methods return clean Python dataclasses: ```python @dataclass class UnifiedMarket: market_id: str # Use this for create_order title: str outcomes: List[MarketOutcome] volume_24h: float liquidity: float url: str # ... more fields @dataclass class MarketOutcome: outcome_id: str # Use this for fetch_ohlcv/fetch_order_book/fetch_trades label: str # "Trump", "Yes", etc. price: float # 0.0 to 1.0 (probability) # ... more fields ``` See the [full API reference](../../API_REFERENCE.md) for complete documentation. ## Important Notes ### Use `outcome.outcome_id`, not `market.market_id` For deep-dive methods like `fetch_ohlcv()`, `fetch_order_book()`, and `fetch_trades()`, you must use the **outcome ID**, not the market ID: ```python markets = poly.fetch_markets(query="Trump") outcome_id = markets[0].outcomes[0].outcome_id # Correct candles = poly.fetch_ohlcv(outcome_id, ...) # Works candles = poly.fetch_ohlcv(markets[0].market_id, ...) # Wrong! ``` ### Prices are 0.0 to 1.0 All prices represent probabilities (0.0 to 1.0). Multiply by 100 for percentages: ```python outcome = markets[0].outcomes[0] print(f"Price: {outcome.price * 100:.1f}%") # "Price: 55.3%" ``` ### Timestamps are Unix milliseconds ```python from datetime import datetime candle = candles[0] dt = datetime.fromtimestamp(candle.timestamp / 1000) print(dt) ``` ## Development ```bash # Clone the repo git clone https://github.com/qoery-com/pmxt.git cd pmxt/sdks/python # Install in development mode pip install -e ".[dev]" # Run tests pytest ``` ## License MIT
text/markdown
PMXT Contributors
null
null
null
MIT
prediction markets, polymarket, kalshi, trading, api
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Pyt...
[]
null
null
>=3.8
[]
[]
[]
[ "urllib3>=1.26.0", "python-dateutil>=2.8.0", "pydantic>=2.0.0", "typing-extensions>=4.0.0", "pytest>=7.0.0; extra == \"dev\"", "pytest-asyncio>=0.21.0; extra == \"dev\"", "black>=23.0.0; extra == \"dev\"", "mypy>=1.0.0; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://github.com/pmxt-dev/pmxt", "Documentation, https://github.com/pmxt-dev/pmxt#readme", "Repository, https://github.com/pmxt-dev/pmxt", "Issues, https://github.com/pmxt-dev/pmxt/issues" ]
twine/6.2.0 CPython/3.14.3
2026-02-19T11:57:35.239338
pmxt-2.10.0.tar.gz
2,023,822
69/7a/c44792a1e0831e1d17c042399280e769f6417906e71764b044dd82934504/pmxt-2.10.0.tar.gz
source
sdist
null
false
4cacc373e83955e9f38e6289c9a8b8c8
bf95aa3717bd0fccc270f07d96b9fa57b0bd80c4cd76b06b5c122cc2c1ab103a
697ac44792a1e0831e1d17c042399280e769f6417906e71764b044dd82934504
null
[]
794
2.4
figpack
0.3.17
A Python package for creating shareable, interactive visualizations in the browser
# figpack [![Tests](https://github.com/flatironinstitute/figpack/actions/workflows/test.yml/badge.svg)](https://github.com/flatironinstitute/figpack/actions/workflows/test.yml) [![codecov](https://codecov.io/gh/flatironinstitute/figpack/branch/main/graph/badge.svg)](https://codecov.io/gh/flatironinstitute/figpack) [![PyPI version](https://badge.fury.io/py/figpack.svg)](https://badge.fury.io/py/figpack) A Python package for creating shareable, interactive visualizations in the browser. ## Documentation For detailed guidance, tutorials, and API reference, visit our **[documentation](https://flatironinstitute.github.io/figpack)**. ## Quick Start Want to jump right in? Here's how to get started: ```bash pip install figpack ``` ```python import numpy as np import figpack.views as vv # Create a timeseries graph graph = vv.TimeseriesGraph(y_label="Signal") # Add some data t = np.linspace(0, 10, 1000) y = np.sin(2 * np.pi * t) graph.add_line_series(name="sine wave", t=t, y=y, color="blue") # Display the visualization in your browser graph.show(open_in_browser=True, title="Quick Start Example") ``` ## License Apache-2.0 ## Citation If you use figpack in your research, please cite it: [![DOI](https://zenodo.org/badge/DOI/10.5281/zenodo.17419621.svg)](https://doi.org/10.5281/zenodo.17419621) ```bibtex @software{magland_figpack_2025, author = {Magland, Jeremy}, title = {figpack}, year = 2025, publisher = {Zenodo}, doi = {10.5281/zenodo.17419621}, url = {https://doi.org/10.5281/zenodo.17419621} } ``` Or in APA format: > Magland, J. (2025). figpack (Version 0.14) [Computer software]. Zenodo. https://doi.org/10.5281/zenodo.17419621 ## Contributing Visit the [GitHub repository](https://github.com/flatironinstitute/figpack) for issues, contributions, and the latest updates.
text/markdown
null
Jeremy Magland <jmagland@flatironinstitute.org>
null
null
Apache-2.0
visualization, plotting, timeseries, interactive
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.8", "Programming Langua...
[]
null
null
>=3.8
[]
[]
[]
[ "numpy", "zarr", "requests", "psutil", "pytest>=7.0; extra == \"test\"", "pytest-cov>=4.0; extra == \"test\"", "pytest-mock>=3.10; extra == \"test\"", "spikeinterface; extra == \"test\"", "matplotlib; extra == \"test\"", "plotly; extra == \"test\"", "Pillow; extra == \"test\"", "pandas; extra ...
[]
[]
[]
[ "Homepage, https://github.com/flatironinstitute/figpack", "Repository, https://github.com/flatironinstitute/figpack", "Documentation, https://flatironinstitute.github.io/figpack", "Bug Tracker, https://github.com/flatironinstitute/figpack/issues" ]
twine/6.1.0 CPython/3.13.7
2026-02-19T11:57:21.020464
figpack-0.3.17.tar.gz
845,093
57/d7/7ada267fb519e0728815fcf818257af277203d5c3e3b4d1bfe795f55284b/figpack-0.3.17.tar.gz
source
sdist
null
false
863fb0dc6250cf235732ddd012653bee
27297c4ba2ef7e70c73cc9d05821582bdc51758096dcb5459e856c1b6649662d
57d77ada267fb519e0728815fcf818257af277203d5c3e3b4d1bfe795f55284b
null
[ "LICENSE" ]
387
2.4
phederation
0.4.4
Phederation implementation of ActivityPub. Can be used to setup a server in the fediverse, and can also be used as a library to build clients connecting to these (and other ActivityPub) servers.
A python/fastAPI implementation of the ActivityPub protocol. Requirements ============ - Linux environment (tested on Mint/Ubuntu) - Python (3.10+) Installation ============ 0) As a pip package: If you want to build ActivityPub clients, you can use the software as a library. .. code-block:: sh pip install phederation 1) For development of the library, clone the repository from codeberg: .. code-block:: sh git clone https://codeberg.org/feldie/phederation.git phederation 2) Navigate to the root folder of the "phederation" repository, then set up a python environment, e.g.: .. code-block:: sh cd phederation python3 -m venv .venv 3) You can start a debug/development version using: .. code-block:: sh source .venv/bin/activate gunicorn -c=settings/gunicorn.conf.dev.py This will start a local server (localhost), you can try to access the fastapi doc at "<localhost:port>/docs". Testing and profiling ===================== There are unit tests in the tests/ folder. They can be run using pytest: .. code-block:: sh pytest tests/ For coverage: .. code-block:: sh coverage run -m pytest tests/ coverage html Profiling can be done with the pytest plugin `pytest-profiling`. For pretty svg plots of the runtimes, we use the additional graphviz tool. .. code-block:: sh pip install pytest-profiling sudo apt install graphviz pytest tests/unit_tests/ --profile-svg Production environment ====================== For a production environment, you should set up proper web hosting. Minimal safeguards should be in place, at least proxy+load balancing+ddos safe; nginx+fail2ban are a good combination: - WSGI server (gunicorn: https://gunicorn.org/) - Proxy and load balancer (nginx: https://nginx.org) - ddos safeguard (fail2ban: https://github.com/fail2ban/fail2ban/wiki) You should also use a proper WSGI webserver (e.g. gunicorn), see below. Phederation is set up to accommodate all of this, and is tested to work in a production environment. Example settings files for production can be found in the "settings" folder. After setting up fail2ban and nginx, you can start a production mode version like this (including creating a logs folder if it does not exist yet): .. code-block:: sh source .venv/bin/activate mkdir logs gunicorn -c=settings/gunicorn.conf.prod.py The command above starts gunicorn workers and serves content on "127.0.0.1:8082" (localhost, then use proxy to serve on the web). It uses the settings file "settings/instance_config-prod.yaml" for production mode. Of course, it is useful to convert this into a proper service that runs continuously. This can be done by containerization of the phederation server itself, building a docker image and then hosting it. It is important to first set the proper variables in the file `settings/.env` (an example file with dummy values is located in `settings/.env.noignore`), then build the docker image, and then start all docker images. Note that the docker compose file `settings/compose.yaml` will start a keycloak server, its postgres database, a mongo database for the phederation image, and the phederation image itself. If you have another keycloak server running already, you should disable the former in the `settings/compose.yaml` file and instead link to yours in the `settings/.env` file. .. code-block:: sh docker build -t phederation-0.1 docker compose up -d
text/x-rst
Felix Dietrich
null
null
null
null
null
[ "Development Status :: 3 - Alpha", "Programming Language :: Python :: 3.12", "Framework :: Pydantic :: 2", "Framework :: FastAPI", "Operating System :: OS Independent", "Intended Audience :: Science/Research", "Topic :: Communications", "Topic :: Software Development :: Libraries :: Python Modules", ...
[]
null
null
>=3.12
[]
[]
[]
[ "aiofiles>=24.1.0", "build>=1.4.0", "coverage>=7.10.7", "cryptography>=45.0.5", "fastapi[standard]==0.124.4", "gunicorn[gevent]>=23.0.0", "httpx>=0.27.2", "mkdocs>=1.6.1", "mkdocs-material>=9.7.1", "mkdocstrings[python]>=1.0.0", "psycopg2-binary>=2.9.10", "pydantic>=2.11.7", "pydantic-settin...
[]
[]
[]
[ "Homepage, https://codeberg.org/feldie/phederation", "Issues, https://codeberg.org/feldie/phederation/-/issues" ]
twine/6.2.0 CPython/3.12.12
2026-02-19T11:56:53.506477
phederation-0.4.4.tar.gz
134,044
06/fe/0fc1f3fdc946f18058d9e4907c8477a5add18503bbc60edb16e9aabf1701/phederation-0.4.4.tar.gz
source
sdist
null
false
e0feb370d4f831b873de264e0cbe5741
f64e68487f2fe4441f6f2d205e3c89890c151973fe6e34e14e808c11d087bf4d
06fe0fc1f3fdc946f18058d9e4907c8477a5add18503bbc60edb16e9aabf1701
MIT
[ "CONTRIBUTING", "LICENSE" ]
236
2.4
humenv
0.0.1
A professional-grade AI utility for automated data synchronization and backend management.
# Installation To install requirements: `python -m pip install requirements.txt` To save requirements: `python -m pip list --format=freeze --exclude-editable -f https://download.pytorch.org/whl/torch_stable.html > requirements.txt` * Note we use Python 3.9.4 for our experiments # Running the code For remaining experiments: Navigate to the corresponding directory, then execute: `python run.py -m` with the corresponding `config.yaml` file (which stores experiment configs). # License Consult License.md
text/markdown
null
AI Research Team <Ai-model@example.com>
null
null
null
automation, api-client, sync, tooling
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent" ]
[]
null
null
>=3.8
[]
[]
[]
[ "requests>=2.28.0", "urllib3>=1.26.0" ]
[]
[]
[]
[ "Homepage, https://github.com/ai/library", "Bug Tracker, https://github.com/ai/library/issues" ]
twine/6.2.0 CPython/3.14.3
2026-02-19T11:56:50.390046
humenv-0.0.1.tar.gz
3,542
47/72/4135d455bd722a665593385087fba77ef798435d7acc49c8176baebeddb4/humenv-0.0.1.tar.gz
source
sdist
null
false
6ec9d5e55628760dd23efd5711bb39c0
b91711ba6bbe3b130613303fbc03f17db669c0cd00b1ae585bb8160c38e78f8c
47724135d455bd722a665593385087fba77ef798435d7acc49c8176baebeddb4
null
[ "LICENSE.txt" ]
258
2.4
iriusrisk-cli
0.6.2
AI-powered threat modeling integration for IriusRisk. Command line interface and MCP server for security analysis.
# IriusRisk CLI An AI-powered threat modeling integration that brings IriusRisk security analysis directly into your development workflow. Designed primarily for use with AI-enabled IDEs through MCP (Model Context Protocol), this tool enables AI assistants to help you create threat models, analyze security risks, and implement countermeasures seamlessly within your coding environment. > :warning: This software has been released IriusRisk Labs as public beta. It is provided as-is, without warranty or guarantee of any kind. Features may change, data structures may evolve, and occasional issues should be expected. Use at your own discretion. > 🎉 **What's New in v0.6.0**: Agent Skills and MCP Tool Filtering! All workflow prompts converted to portable **Agent Skills** following the [Agent Skills open standard](https://agentskills.io) - now works across any agent (Cursor, Windsurf, Cline, etc.). Skills organized by LLM capability (reasoning, general, code-focused). Plus **MCP tool filtering** for fine-grained control over which tools are exposed to AI clients. See [CHANGELOG.md](CHANGELOG.md) for details. ## Primary Use Case: AI-Enabled IDE Integration This tool is designed to work alongside AI assistants in your IDE, enabling: - **AI-Guided Threat Modeling**: Let AI assistants analyze your code and automatically create comprehensive threat models - **Intelligent Security Analysis**: Get AI-powered insights on threats and countermeasures specific to your codebase - **Contextual Security Recommendations**: Receive security guidance based on your actual code changes and architecture - **Automated Security Workflows**: Have AI assistants track threat status, implement countermeasures, and generate reports ## What You Can Do - **Agent Skills**: Portable AI workflow skills that work across any compatible agent (Cursor, Windsurf, Cline, etc.), organized by LLM capability - **MCP Tool Filtering**: Fine-grained control over which MCP tools are exposed to AI clients, with category-based and per-tool filtering - **CI/CD Security Verification**: Automated security drift detection in CI/CD pipelines with three specialized tools: - Compare threat model versions to detect architectural and security changes - Verify security control implementations in code - Generate comprehensive security reports for PR reviews and deployments - **Multi-Repository Threat Models**: Unite security across microservices, infrastructure, and frontend repos into one comprehensive view - **AI-Guided Threat Modeling**: Let AI assistants analyze your code and automatically create threat models - **Import/Export Threat Models**: Use OTM (Open Threat Model) format to create and update projects, with automatic schema validation - **Intelligent Questionnaires**: AI analyzes your code to answer security questionnaires, refining threat models based on actual implementation - **Analyze Threats**: View, filter, search, and update threat status with scope-based filtering for multi-repo projects - **Track Countermeasures**: Monitor implementation, update status, and create tracking issues - **Generate Reports**: Create compliance and security reports in multiple formats (PDF, HTML, XLSX, CSV) - **Version Control**: Automatic snapshots of your threat models to track changes over time - **Issue Tracker Integration**: Connect countermeasures to your issue tracking system - **MCP Integration**: Enable AI assistants to perform all operations through Model Context Protocol ## Key Features ### 🆕 Agent Skills - Portable AI Workflows (v0.6.0) All workflow prompts are now available as portable **Agent Skills** following the [Agent Skills open standard](https://agentskills.io). Skills work across any agent that supports the standard (Cursor, Windsurf, Cline, etc.) and are organized by LLM capability—reasoning, general, and code-focused. Version-controlled and shareable across projects and teams. See the [Agent Skills](#agent-skills---portable-ai-workflows) section for the full skill catalog and usage instructions. ### 🆕 MCP Tool Filtering (v0.6.0) Control which MCP tools are exposed to AI clients by filtering based on categories or specific tool names. Reduce scope for specific clients, use custom workflow instructions instead of built-in guidance, or create specialized MCP configurations for different roles and teams. See the [Filtering MCP Tools](#filtering-mcp-tools) section for configuration details. ### CI/CD Security Verification (v0.5.0) Automated security drift detection and control verification for CI/CD pipelines. Three specialized MCP tools enable AI-powered security reviews: version comparison for drift detection, countermeasure verification against code, and comprehensive orchestrated security gates for pre-deployment checks. See the [CI/CD Security Verification](#cicd-security-verification-v050) section for detailed workflows and examples. ### Multi-Repository Threat Modeling (v0.4.0) Multiple repositories can now contribute to a single unified threat model—perfect for microservices, infrastructure-as-code, and distributed architectures. Each repository defines its **scope** (what it represents), and the AI intelligently merges contributions into a comprehensive security view. **Key benefits**: - Unified security view across all repositories - Each team maintains their own codebase while contributing to shared threat model - AI-driven intelligent merging based on scope definitions - Scope-based filtering shows only relevant threats per repository See the [Multi-Repository Threat Modeling](#multi-repository-threat-modeling) section for detailed examples and workflows. ### Questionnaire-Driven Threat Model Refinement (v0.3.0) IriusRisk questionnaires help refine your threat model based on your actual implementation details. The CLI downloads questionnaires during sync operations, and when using the MCP integration, AI assistants can analyze your source code to automatically answer these questions. Your answers trigger IriusRisk's rules engine to regenerate your threat model with more accurate, context-specific threats and countermeasures. **How it works**: 1. Import or create your initial threat model 2. Run `iriusrisk sync` to download questionnaires 3. AI assistant analyzes your code and answers the questions 4. Sync automatically pushes answers back to IriusRisk 5. IriusRisk regenerates your threat model based on actual implementation ### AI-Powered Security Analysis Through MCP integration, AI assistants can analyze your codebase, create threat models, answer questionnaires, implement countermeasures, and track progress—all within your IDE. The AI understands your code context and provides intelligent, specific security guidance. ### Automatic Threat Model Regeneration The CLI automatically triggers IriusRisk's rules engine to regenerate your threat model when needed—after answering questionnaires, making manual edits, or updating your architecture. This happens seamlessly during sync operations without manual intervention. ### Enhanced Auto-Versioning The CLI automatically creates backup snapshots **before** any update operation. Roll back changes anytime and track your threat model's evolution over time. Works consistently across all import scenarios (CLI commands, MCP tools, project updates). # MCP Integration for AI-Enabled IDEs ## Setting Up MCP Integration The IriusRisk CLI is designed to work with AI assistants through MCP (Model Context Protocol). This enables your AI assistant to: - Analyze your codebase and create threat models automatically - Provide security recommendations based on your specific code - Track and update threat and countermeasure status - Generate security reports and documentation ### Configuration for MCP 1. Install the IriusRisk CLI (see installation instructions below) 2. Configure your IriusRisk connection with environment variables 3. Your AI assistant will automatically detect and use the MCP integration ### AI Assistant Capabilities When integrated through MCP, AI assistants can: - **Analyze Source Code**: Examine your application code, infrastructure, and documentation to identify security-relevant components - **Create Threat Models**: Generate comprehensive OTM (Open Threat Model) files from your codebase - **Import to IriusRisk**: Automatically upload threat models to IriusRisk for professional analysis - **Answer Questionnaires**: Analyze your code to answer project and component questionnaires, refining the threat model based on actual implementation details - **Review Threats**: Help you understand and prioritize security threats identified by IriusRisk - **Implement Countermeasures**: Guide you through implementing security controls and track their status - **Sync Changes**: Automatically synchronize threat model updates between your local environment and IriusRisk - **Generate Reports**: Create compliance reports and security documentation ## Agent Skills - Portable AI Workflows **New in v0.6.0**: All workflow prompts have been converted to portable **Agent Skills** following the [Agent Skills open standard](https://agentskills.io). Agent Skills are portable, version-controlled packages that teach AI agents how to perform domain-specific tasks. They work across any agent that supports the standard (Cursor, Windsurf, Cline, etc.) and can be shared, versioned, and reused. ### Skills by Category Skills are organized by LLM capability to help you choose the right model for each task: #### Reasoning Models (Complex Analysis) Best for: Claude Sonnet, GPT-4, o1, other reasoning-capable models - **ci-cd-verification** - Orchestrate comprehensive CI/CD security reviews - **compare-versions** - Compare threat model versions, interpret structured diffs - **countermeasure-verification** - Verify security controls are correctly implemented in code #### General Models (Standard Workflows) Best for: Most modern LLMs (Claude, GPT-4, Gemini, etc.) - **architecture-design-review** - Architecture/design review trigger point - **initialize-iriusrisk-workflow** - Complete workflow instructions for all IriusRisk operations - **analyze-source-material** - Analyze repositories to extract components for threat modeling - **create-threat-model** - Step-by-step OTM file creation with validation - **threats-and-countermeasures** - Analyze threats/countermeasures, provide implementation guidance - **security-development-advisor** - Help developers assess security impact #### Code-Focused (Heavy Analysis) Best for: Models with strong code analysis capabilities - **questionnaire-guidance** - Analyze source code to answer IriusRisk questionnaires #### Shared (Reference Material) Best for: All models - **otm-layout-guidance** - Detailed OTM component layout and positioning guidance - **otm-validation-guidance** - Validation rules for trust zones and component types ### Using Agent Skills In Cursor and other compatible agents, skills are automatically discovered from the `skills/` directory in your project or globally from `~/.cursor/skills/`. **To use a skill:** 1. Type `/` in Agent chat 2. Search for the skill name 3. Or let the agent automatically apply it based on context **Learn more:** See [skills/README.md](skills/README.md) for detailed documentation on each skill and model recommendations. ## Example AI Workflows ### Single Repository 1. **Code Analysis**: "Analyze my web application for security threats" 2. **Threat Model Creation**: AI examines your code and creates a comprehensive threat model 3. **IriusRisk Integration**: Threat model is uploaded to IriusRisk for professional analysis 4. **Questionnaire Completion**: AI analyzes your code to answer questionnaires, refining the threat model 5. **Threat Review**: AI helps you understand and prioritize identified threats 6. **Implementation Guidance**: AI guides you through implementing security countermeasures 7. **Status Tracking**: Progress is tracked and synchronized with IriusRisk ### Multi-Repository 1. **First Repo (Backend)**: "Threat model this API" → AI creates initial threat model with API components 2. **Second Repo (Infrastructure)**: "Update threat model with AWS infrastructure" → AI exports existing model, adds infrastructure, shows API running in ECS 3. **Third Repo (Frontend)**: "Add frontend to threat model" → AI merges frontend components, creates complete system view 4. **Result**: Single comprehensive threat model showing frontend → load balancer → API → database with all security threats across the stack ## Multi-Repository Threat Modeling Modern applications are often distributed across multiple repositories (microservices, infrastructure-as-code, frontend/backend separation). The IriusRisk CLI supports **scope definitions** that enable multiple repositories to contribute different perspectives to a single comprehensive threat model. ### How It Works Each repository defines its **scope**—a description of what part of the system it represents. The AI assistant then exports any existing threat model, analyzes the current repository based on its scope, intelligently merges contributions with existing components, and updates the unified threat model in IriusRisk. ### Example: Microservices Architecture ```bash # Repository 1: Backend API $ cd backend-api/ $ iriusrisk init -n "E-commerce Platform" --scope "Node.js REST API for order processing..." # AI creates initial threat model with API components # Repository 2: Infrastructure $ cd ../terraform-aws/ $ iriusrisk init -r "ecommerce-platform" --scope "AWS infrastructure - ECS, RDS, VPC..." # AI exports existing model, adds AWS infrastructure, merges intelligently # Repository 3: Frontend $ cd ../web-frontend/ $ iriusrisk init -r "ecommerce-platform" --scope "React SPA for customer interface..." # AI exports existing model, adds frontend components, creates complete view ``` **Result**: A single, comprehensive threat model showing frontend → infrastructure → backend services and how they all connect and interact, with complete security coverage across the entire stack. ### Defining Scope A good scope description includes what this repository contains, how it relates to other repositories, where components are deployed, external integrations, and trust boundaries. **Example scope definitions:** ```json { "scope": "Node.js REST API implementing core business logic for order processing, inventory management, and user accounts. Exposes REST endpoints consumed by web frontend (separate repo) and mobile apps. Connects to PostgreSQL database and Redis cache. Deployed as containerized service (infrastructure in terraform repo). Integrates with external payment gateway (Stripe) and shipping APIs." } ``` ```json { "scope": "AWS cloud infrastructure via Terraform. Provisions ECS Fargate containers for the API backend (api-backend repo), RDS PostgreSQL database, ElastiCache Redis, Application Load Balancer, CloudFront CDN for frontend static assets (web-frontend repo), S3 buckets, and VPC with public/private subnet isolation. All application components from other repos run within these AWS services." } ``` ### Setting Up Multi-Repository Projects **First repository (creates project):** ```bash $ cd backend-api/ $ iriusrisk init -n "E-commerce Platform" # Optional: Add scope when prompted, or skip for complete system view ``` **Additional repositories (connect to existing project):** ```bash $ cd ../terraform-aws/ $ iriusrisk init -r "ecommerce-platform" # Prompted for scope to define this repository's contribution ``` Or use the `--scope` flag to skip prompts: ```bash $ iriusrisk init -r "ecommerce-platform" --scope "AWS infrastructure via Terraform..." ``` ### Working with Multi-Repository Projects When you ask the AI to create or update a threat model, it automatically runs `sync()` to download the latest threat model, analyzes your code based on the repository's scope, merges your contribution with existing components, and updates the unified threat model in IriusRisk. The `sync()` command saves the current model to `.iriusrisk/current-threat-model.otm`, making it immediately available for intelligent merging. **Important:** The AI creates ALL OTM files in the `.iriusrisk/` directory with clear temporary naming (e.g., `.iriusrisk/temp-update-20260206-143022.otm`). It ALWAYS runs `sync()` first to download the current state, then uses IDENTICAL merge logic whether updating a single repository or merging multi-repository contributions. This ensures: - Changes made in IriusRisk (questionnaires, threat status, countermeasures) are never overwritten - Existing component layout positions are preserved - All updates follow the same reliable merge algorithm ### Scope-Based Filtering AI assistants automatically filter threats, countermeasures, and questionnaires to show only items relevant to your repository's scope. Infrastructure repos see infrastructure threats, application repos see application-level threats, and frontend repos see client-side threats. Each repository stays focused on its own security concerns while contributing to the unified threat model. ## CI/CD Security Verification (v0.5.0) Automated security drift detection and control verification for CI/CD pipelines. Three specialized MCP tools enable AI-powered security reviews at different stages of your development workflow. ### The Three Tools **1. `compare_versions` - Version Comparison** Compare any two threat model versions to detect architectural and security changes. Returns structured JSON diff showing what changed. ``` # Simple drift detection "Compare baseline version to current" # Historical audit "Compare v1.0-approved to v2.0-approved" ``` **Use when:** Checking for security drift, auditing changes, understanding threat model evolution. **2. `countermeasure_verification` - Control Implementation Verification** Verify that security controls linked to issue tracker tasks (Jira, GitHub Issues, etc.) are correctly implemented in code. ``` # Verify specific issue "Verify that PROJ-1234 is correctly implemented" # Verify all linked controls in PR "Check if security controls are properly implemented" ``` **Use when:** Reviewing PRs that claim to fix security issues, validating control implementations, ensuring countermeasures match code. **3. `ci_cd_verification` - Comprehensive Security Review** Orchestrator that coordinates complete security reviews by combining version comparison, control verification, risk analysis, and reporting. ``` # Full CI/CD security gate "Run comprehensive security verification against baseline" ``` **Use when:** CI/CD pipelines, pre-deployment security gates, comprehensive PR reviews requiring multiple analyses. ### Headless Automation Script For CI/CD pipelines, use the included `iriusrisk-verification.sh` script: ```bash # Run from your project directory ./iriusrisk-verification.sh --type <verification-type> # Types: # drift - Detect drift from baseline # pr - Full PR security review # control - Verify control implementation # comprehensive - Complete security gate # audit - Historical version comparison ``` **Example CI/CD Integration:** ```yaml # .github/workflows/security-check.yml name: Security Verification on: [pull_request] jobs: security: runs-on: ubuntu-latest steps: - uses: actions/checkout@v2 - name: Run Security Verification run: | ./iriusrisk-verification.sh --type pr env: ANTHROPIC_API_KEY: ${{ secrets.ANTHROPIC_API_KEY }} IRIUSRISK_API_TOKEN: ${{ secrets.IRIUSRISK_API_TOKEN }} IRIUSRISK_DOMAIN: release.iriusrisk.com ``` ### What Gets Detected **Architectural Changes:** - Components added, removed, or modified - Dataflows added, removed, or modified (including trust boundary crossings) - Trust zone changes (components moved between security zones) **Security Changes:** - New threats introduced (with severity levels) - Threat severity increases - Countermeasures added or removed - Critical security control removals **Control Implementation:** - Whether security controls are correctly implemented in code - Evidence-based verification with specific file/line references - Pass/fail status with detailed reasoning ### Workflow Examples **Drift Detection:** ``` > Compare baseline version to current state AI will: 1. Compare versions using compare_versions tool 2. Identify architectural and security changes 3. Generate DRIFT_REPORT.md with findings ``` **PR Security Review:** ``` > Run security review of this pull request AI will: 1. Analyze current code and create updated OTM 2. Import OTM to IriusRisk 3. Compare against baseline version 4. Generate PR_REPORT.md with security assessment ``` **Control Verification:** ``` > Verify that PROJ-1234 is correctly implemented AI will: 1. Find countermeasure linked to PROJ-1234 2. Analyze code changes 3. Verify implementation matches requirement 4. Generate CONTROL_VERIFICATION_REPORT.md with pass/fail status ``` # Using the CLI ## Installation ### From PyPI Eventually users will be able to install the CLI using: ```bash $ pip install iriusrisk-cli ``` For Mac users using Homebrew, we suggest installing it with: ```bash $ pipx install iriusrisk-cli ``` ### For Development Clone this repository and install in development mode: ```bash $ git clone <repository-url> $ cd iriusrisk_cli $ pip install -e . ``` ## Configuration Before using the CLI, you need to configure your IriusRisk connection. The CLI supports multiple configuration methods with a clear priority order. ### Recommended: User-Level Configuration Set up your credentials once for use across all projects: ```bash # Set your default IriusRisk hostname iriusrisk config set-hostname https://your-instance.iriusrisk.com # Set your API key (prompts securely, not stored in shell history) iriusrisk config set-api-key # View your current configuration iriusrisk config show ``` This approach: - Keeps your API key secure (not in project files) - Works across all projects automatically - Can be overridden per-project or per-session ### Configuration Priority The CLI checks configuration sources in this order (highest to lowest): 1. **Environment variables** - `IRIUS_HOSTNAME` and `IRIUS_API_KEY` (or `IRIUS_API_TOKEN`) 2. **Project .env file** - `.env` in your project directory 3. **Project config** - `.iriusrisk/project.json` (hostname only, never API credentials) 4. **User config** - `~/.iriusrisk/config.json` (set via `iriusrisk config` commands) Each setting is resolved independently, so you can mix sources (e.g., API key from user config, hostname from environment variable). ### Alternative Configuration Methods #### Option 2: Project .env file Create a `.env` file in your project directory: ```bash cat > .env << EOF IRIUS_HOSTNAME=https://your-instance.iriusrisk.com IRIUS_API_KEY=your-api-token-here EOF ``` **Warning**: If using `.env` files, add them to `.gitignore` to avoid committing credentials. #### Option 3: Environment variables ```bash export IRIUS_HOSTNAME=https://your-instance.iriusrisk.com export IRIUS_API_KEY=your-api-token-here ``` #### Option 4: Project-specific hostname For teams working with different IriusRisk instances, you can set a hostname in the project config: ```bash # Manually edit .iriusrisk/project.json and add: { "hostname": "https://dev-instance.iriusrisk.com", "project_id": "...", ... } ``` **Note**: API credentials should never be stored in project config files. ## Logging and Output Control The IriusRisk CLI provides flexible logging options to control output verbosity: ### Default Behavior By default, the CLI operates quietly with minimal output - only showing command results and critical errors. ### Logging Options ```bash # Enable verbose output (shows progress and status messages) iriusrisk --verbose project list # Enable debug output (shows detailed API calls and timing) iriusrisk --debug project list # Suppress all non-essential output (quiet mode) iriusrisk --quiet project list # Write logs to a specific file iriusrisk --log-file debug.log --debug project list # Set specific log level iriusrisk --log-level INFO project list ``` ### Environment Variables You can also control logging through environment variables: ```bash # Enable debug mode export IRIUSRISK_DEBUG=1 # Set log file path export IRIUSRISK_LOG_FILE=debug.log ``` ### Output Destinations - **stdout**: Command results and data (for piping/redirection) - **stderr**: Status messages, progress, warnings, errors, debug info - **Log files**: Only when explicitly requested via `--log-file` ## Testing API Connection After configuration, test your connection to ensure everything is working correctly: ```bash # Test your IriusRisk connection iriusrisk test ``` This command will: - Test connectivity to your IriusRisk instance - Verify your authentication credentials - Display your IriusRisk version information Example output: ``` Testing connection to IriusRisk... ✓ Connection successful! ✓ IriusRisk version: 4.47.19-0-g41bcb27de1f-30/09/2025 17:48 ``` If the test fails, it will provide specific error information to help you troubleshoot configuration issues. ## Getting help Users can get help using the following commands: ```bash $ iriusrisk help # Detailed help with examples and configuration $ iriusrisk --help # Basic command help $ iriusrisk --version # Show version information ``` ## Quick Start After installation and configuration: ```bash # Test the installation $ iriusrisk --version # Test your API connection $ iriusrisk test # Get detailed help $ iriusrisk help # Basic help $ iriusrisk --help # List projects $ iriusrisk project list # List projects with filtering $ iriusrisk project list --name "web" --format json ``` ## Available Commands ### Project Initialization Initialize a new project or connect to an existing one: ```bash # Initialize new project interactively $ iriusrisk init # Initialize with specific name $ iriusrisk init -n "My Web Application" # Initialize with name and project ID $ iriusrisk init -n "My App" -p abc123 # Connect to existing project by reference ID $ iriusrisk init -r "my-project-ref" # Connect to existing project with scope definition (multi-repository) $ iriusrisk init -r "my-project-ref" --scope "AWS infrastructure via Terraform" # Initialize with scope for multi-repository project $ iriusrisk init -n "My App" --scope "Backend API services and business logic" # Overwrite existing configuration $ iriusrisk init --force ``` **Multi-Repository Projects**: Use the `--scope` parameter to define how this repository contributes to a unified threat model. The scope helps AI assistants merge contributions from multiple repositories. See the Multi-Repository Threat Modeling section above for details. ### Projects ```bash # List all projects $ iriusrisk project list # List projects with pagination $ iriusrisk project list --page 1 --size 10 # Filter by name (partial match) $ iriusrisk project list --name "web application" # Filter by tags $ iriusrisk project list --tags "production critical" # Filter by workflow state $ iriusrisk project list --workflow-state "in-progress" # Show only non-archived projects $ iriusrisk project list --not-archived # Include version information $ iriusrisk project list --include-versions # Output as JSON $ iriusrisk project list --format json # Output as CSV $ iriusrisk project list --format csv # Advanced filtering with custom expressions $ iriusrisk project list --filter "'name'~'web':AND:'tags'~'prod'" # Show detailed project information $ iriusrisk project show <project_id> # Show project info as JSON $ iriusrisk project show <project_id> --format json ``` ### OTM (Open Threat Model) Import/Export Work with threat models using the standard OTM format: ```bash # Generate example OTM file for reference $ iriusrisk otm example # Import OTM file (automatically creates new or updates existing project) $ iriusrisk otm import example.otm # Import with JSON output $ iriusrisk otm import example.otm --format json # Import and reset diagram layout (forces IriusRisk auto-layout) $ iriusrisk otm import example.otm --reset-layout # Export project as OTM format $ iriusrisk otm export PROJECT_ID # Export to specific file $ iriusrisk otm export PROJECT_ID -o threat-model.otm # Export as JSON $ iriusrisk otm export PROJECT_ID --format json # Export existing threat model for multi-repository merging $ iriusrisk otm export -o existing-model.otm ``` **Note**: The CLI automatically detects whether you're creating a new project or updating an existing one during import. If auto-versioning is enabled in your project configuration, a backup snapshot is automatically created before updates. #### OTM Schema Validation All OTM files are automatically validated against the official [Open Threat Model JSON schema](https://github.com/iriusrisk/OpenThreatModel) before import. This catches structural issues early and prevents data loss: ```bash $ iriusrisk otm import threat-model.otm 🔍 Validating OTM file against schema... ✓ OTM validation passed Importing OTM file: threat-model.otm ... ``` If validation fails, you'll see clear error messages: ``` ❌ OTM validation failed! Validation errors: • At 'project': 'id' is a required property • At 'components -> 0': 'parent' is a required property • At 'dataflows -> 0 -> source': 'component-xyz' does not exist OTM file summary: Project: My App (ID: None) Trust Zones: 2 Components: 5 Dataflows: 3 Threats: 0 Mitigations: 0 ⚠️ Please fix the validation errors before importing. See OTM specification: https://github.com/iriusrisk/OpenThreatModel ``` **What validation catches:** - Missing required fields: - `otmVersion` (required at root) - `project.name` and `project.id` (required) - `component.parent` (trustZone or component - required) - `trustZone.risk.trustRating` (required) - `dataflow.source` and `dataflow.destination` (required) - Invalid data types (strings vs numbers, etc.) - Malformed structure - Schema violations **Benefits:** - **Prevents data loss** - Catches issues before they reach IriusRisk - **Clear errors** - Shows exactly what's wrong and where - **Early detection** - Fails fast before API calls - **Spec compliance** - Ensures OTM files follow official specification - **Better AI output** - AI can see validation errors and fix them **Dependencies:** - Requires `jsonschema>=4.0.0` (automatically installed) - Requires `pyyaml>=6.0.0` (automatically installed) #### Layout Reset Feature Use `--reset-layout` to strip all component positions and sizes, forcing IriusRisk to auto-layout the diagram from scratch: ```bash # One-time layout reset $ iriusrisk otm import threat-model.otm --reset-layout ``` Or enable automatic layout reset for all imports in `.iriusrisk/project.json`: ```json { "name": "My Project", "reference_id": "my-project-abc", "auto_versioning": true, "auto_reset_layout": false } ``` **When to use layout reset:** - Diagram has become messy after multiple updates - Major architectural refactoring makes old positions irrelevant - Want IriusRisk's auto-layout to reorganize everything - Testing/debugging with fresh layout **AI usage:** When calling `import_otm()` via MCP, use `reset_layout=True` parameter to reset layout. **Multi-Repository Use**: When contributing to an existing threat model from a new repository, export the current threat model first. This allows AI assistants to merge your contribution with existing components intelligently based on your repository's scope definition. #### OTM File Management Best Practices **All OTM files go in `.iriusrisk/` directory:** - AI creates temporary OTM files with clear naming: `.iriusrisk/temp-update-YYYYMMDD-HHMMSS.otm` - Never creates OTM files in repository root - Temporary files can be deleted after successful import **Always sync first:** - AI ALWAYS runs `iriusrisk sync` before any threat modeling operation - Downloads `.iriusrisk/current-threat-model.otm` with current state from IriusRisk - This is the ONLY authoritative local copy **Identical merge logic for all updates:** - Whether updating a single repository or merging multi-repo contributions - AI uses the SAME merge algorithm: 1. Read `current-threat-model.otm` 2. Preserve ALL existing components, IDs, and layout positions 3. Add NEW components with calculated positions 4. Save to `.iriusrisk/temp-update-YYYYMMDD-HHMMSS.otm` 5. Import the temporary file **Correct workflow:** ```bash # AI workflow (automatic) $ iriusrisk sync # ALWAYS first - downloads current state # AI reads .iriusrisk/current-threat-model.otm # AI merges changes preserving existing components and layout # AI creates .iriusrisk/temp-update-20260206-143022.otm $ iriusrisk otm import .iriusrisk/temp-update-20260206-143022.otm # Manual workflow (if needed) $ iriusrisk sync # Get current state # Manually edit .iriusrisk/current-threat-model.otm # Save as .iriusrisk/temp-update-20260206-143022.otm $ iriusrisk otm import .iriusrisk/temp-update-20260206-143022.otm ``` **What NOT to do:** ```bash # ❌ WRONG - Skip sync # ❌ WRONG - Create OTM files in repo root # ❌ WRONG - Use different logic for single-repo vs multi-repo # ❌ WRONG - Ignore existing layout positions ``` ### Data Synchronization Sync threat model data between IriusRisk and your local environment: ```bash # Sync all data from default project $ iriusrisk sync # Sync specific project $ iriusrisk sync <project_id> # Sync only threats $ iriusrisk sync --threats-only # Sync only countermeasures $ iriusrisk sync --countermeasures-only # Sync only questionnaires $ iriusrisk sync --questionnaires-only # Sync only components $ iriusrisk sync --components-only # Sync to custom output directory $ iriusrisk sync -o /path/to/output ``` The sync command downloads: - **Threats**: All identified security threats for your project - **Countermeasures**: Security controls and their implementation status - **Questionnaires**: Questions to refine your threat model based on implementation details - **Components**: System components from the IriusRisk library and your architecture - **Current Threat Model (OTM)**: The complete threat model exported from IriusRisk for multi-repository merging Data is saved to `.iriusrisk/` directory by default and can be used for offline analysis or AI-assisted review. **Multi-Repository Support**: The `current-threat-model.otm` file enables multiple repositories to merge their contributions intelligently. AI assistants automatically detect and use this file as the basis for updates when contributing from additional repositories. ### Questionnaires **New in 0.3.0**: Questionnaires help refine your threat model based on actual implementation details. ```bash # Download questionnaires during sync $ iriusrisk sync # View questionnaires in .iriusrisk/questionnaires.json ``` Questionnaires are automatically downloaded during sync operations. When using the MCP integration, AI assistants can analyze your source code and automatically answer these questions, triggering IriusRisk's rules engine to regenerate your threat model with more accurate threats and countermeasures. **How it works**: 1. Import or create a threat model 2. Run `iriusrisk sync` to download questionnaires 3. AI assistant analyzes your code to answer questions 4. Sync pushes answers back to IriusRisk 5. IriusRisk automatically regenerates threat model based on answers ### Threats ```bash # List all threats from default project $ iriusrisk threat list # List threats from specific project $ iriusrisk threat list <project_id> # Show detailed threat information $ iriusrisk threat show <threat_id> # Search threats by keyword $ iriusrisk threat search "SQL injection" # Search with project specification $ iriusrisk threat search "XSS" --project-id <project_id> # Update threat status $ iriusrisk threat update <threat_id> --status accept --reason "Mitigated by WAF" # List threats with specific status $ iriusrisk threat list --status required # Output as JSON or CSV $ iriusrisk threat list --format json ``` **Available threat statuses**: `required`, `recommended`, `accept`, `expose`, `not-applicable` ### Countermeasures ```bash # List all countermeasures from default project $ iriusrisk countermeasure list # List countermeasures from specific project $ iriusrisk countermeasure list <project_id> # Show detailed countermeasure information $ iriusrisk countermeasure show <countermeasure_id> # Search countermeasures by keyword $ iriusrisk countermeasure search "authentication" # Update countermeasure status $ iriusrisk countermeasure update <cm_id> --status implemented # Create issue tracker ticket for countermeasure $ iriusrisk countermeasure create-issue <cm_id> # Create issue with specific tracker $ iriusrisk countermeasure create-issue <cm_id> --tracker "Jira" # List countermeasures by status $ iriusrisk countermeasure list --status required # Output as JSON or CSV $ iriusrisk countermeasure list --format json ``` **Available countermeasure statuses**: `required`, `recommended`, `implemented`, `rejected`, `not-applicable` ### Components View and search system components in your architecture: ```bash # List all components $ iriusrisk component list # List components from specific project $ iriusrisk component list <project_id> # Show detailed component information $ iriusrisk component show <component_id> # Search components by keyword $ iriusrisk component search "database" # Filter by category $ iriusrisk component list --category "Database" # Filter by type $ iriusrisk component list --type "project-component" # Output as JSON $ iriusrisk component list --format json ``` ### Reports Generate security and compliance reports: ```bash # Generate default countermeasure report (PDF) $ iriusrisk reports generate # Generate threat report $ iriusrisk reports generate --type threat # Generate compliance report $ iriusrisk reports generate --type compliance --standard owasp-top-10-2021 # Generate report in different format $ iriusrisk reports generate --format html # Save to specific location $ iriusrisk reports generate -o /path/to/report.pdf # List available report types $ iriusrisk reports types # List available compliance standards $ iriusrisk reports standards # List generated reports $ iriusrisk reports list ``` **Available report types**: `countermeasure`, `threat`, `compliance`, `risk-summary` **Available formats**: `pdf`, `html`, `xlsx`, `csv`, `xls` ### Project Versions **New in 0.3.0**: Enhanced auto-versioning creates backup snapshots before updates. ```bash # Create a version snapshot $ iriusrisk project versions create "v1.0" --description "Initial release" # List all versions for a project $ iriusrisk project versions list # List versions for specific project $ iriusrisk project versions list <project_id> # Show version details $ iriusrisk project versions show <version_id> # Compare two versions $ iriusrisk project versions compare <version_id_1> <version_id_2> ``` **Auto-versioning**: Enable in `.iriusrisk/project.json`: ```json { "auto_version": true, "auto_version_prefix": "auto-backup-" } ``` When enabled, the CLI automatically creates version snapshots before OTM imports and updates, protecting your work. ### Updates Tracking Track threat and countermeasure status changes before syncing to IriusRisk: ```bash # View pending updates $ iriusrisk updates show # View updates for specific project $ iriusrisk updates show <
text/markdown
IriusRisk
support@iriusrisk.com
null
null
null
security, threat-modeling, iriusrisk, cli, mcp, ai, threat-analysis, security-testing, compliance, cybersecurity
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "Intended Audience :: Information Technology", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3", "Programming Languag...
[]
https://github.com/iriusrisk/iriusrisk_cli
null
>=3.8
[]
[]
[]
[ "click>=8.0.0", "requests>=2.25.0", "python-dotenv>=0.19.0", "tabulate>=0.8.0", "mcp>=1.0.0", "jsonschema>=4.0.0", "pyyaml>=6.0.0" ]
[]
[]
[]
[ "Bug Reports, https://github.com/iriusrisk/iriusrisk_cli/issues", "Documentation, https://github.com/iriusrisk/iriusrisk_cli#readme", "Source, https://github.com/iriusrisk/iriusrisk_cli", "Changelog, https://github.com/iriusrisk/iriusrisk_cli/blob/main/CHANGELOG.md" ]
twine/6.2.0 CPython/3.14.0
2026-02-19T11:56:11.167157
iriusrisk_cli-0.6.2.tar.gz
265,803
11/37/b46ab356c99c58cd69ef53de0e46132c3b782cdcfebdfc7739f55d98bc98/iriusrisk_cli-0.6.2.tar.gz
source
sdist
null
false
2f47d970faacb95b4d924457d5a61bec
e7dfdeb2988674b36041bd276449b3b9661451f686d7a88e84297021230de223
1137b46ab356c99c58cd69ef53de0e46132c3b782cdcfebdfc7739f55d98bc98
null
[ "LICENSE" ]
250
2.1
nonebot-plugin-rikka
0.4.4
一个简简单单的舞萌机器人插件
<div align=center> <img width=200 src="./assets/RikkaLogo.webp" alt="image"/> <h1 align="center">Nonebot-Plugin-Rikka</h1> <p align="center">一个简单的 NoneBot2 舞萌查询成绩插件</p> </div> <div align=center> <a href="#关于️"><img src="https://img.shields.io/github/stars/Moemu/Nonebot-Plugin-Rikka" alt="Stars"></a> <a href="https://pypi.org/project/Nonebot-Plugin-Rikka/"><img src="https://img.shields.io/pypi/v/Nonebot-Plugin-Rikka" alt="PyPI Version"></a> <a href="https://pypi.org/project/Nonebot-Plugin-Rikka/"><img src="https://img.shields.io/pypi/dm/Nonebot-Plugin-Rikka" alt="PyPI Downloads" ></a> <a href="https://nonebot.dev/"><img src="https://img.shields.io/badge/nonebot-2-red" alt="nonebot2"></a> <a href="#"><img src="https://img.shields.io/badge/Code%20Style-Black-121110.svg" alt="codestyle"></a> </div> > [!NOTE] > > 本项目进入慢更新状态,但是您仍然可以提出新的特性请求 ## 介绍✨ 基于 [Nonebot2](https://nonebot.dev/) 的舞萌DX的查分插件 看板娘: [Rikka](https://bot.snowy.moe/about/Rikka) ## 功能🪄 ✅ 支持游戏: 舞萌DX(Ver.CN 1.53+), ~~中二节奏(Not Plan yet.)~~ ✅ 支持数据源: [落雪咖啡屋(未绑定的首选)](https://maimai.lxns.net/), [水鱼查分器](https://www.diving-fish.com/maimaidx/prober/) ✅ 支持功能: 基础查分功能、拟合系数查询、曲目标签查询 ## 指令列表🕹️ 带有🚧标志的指令暂不可用或仍在开发中 | 指令 | 说明 | | ----------------------------------- | ------------------------------------------------------ | | `.bind lxns\|divingfish` | [查分器相关]绑定游戏账号/查分器 | | `.unbind lxns\|divingfish\|all` | [查分器相关]解绑游戏账号/查分器 | | `.source lxns\|divingfish` | [查分器相关]设置默认查分器 | | `.b50` | [舞萌DX]生成玩家 Best50 | | `.r50` | [舞萌DX]生成玩家 Recent 50(需绑定落雪查分器) | | `.n50` | [舞萌DX]获取玩家拟合系数 Top-50 | | `.ap50` | 🚧[舞萌DX]生成玩家 ALL PERFECT 50 | | `.pc50` | [舞萌DX]生成玩家游玩次数 Top50 | | `.minfo <id\|乐曲名称\|别名>` | [舞萌DX]获取乐曲信息 | | `.random` | [舞萌DX]随机获取一首乐曲 | | `.alias add <song_id> <别名>` | [舞萌DX]添加乐曲别名(不会被 update 操作覆盖) | | `.alias update` | [舞萌DX]从落雪查分器更新乐曲别名数据库 | | `.alias query <id\|乐曲名称\|别名>` | [舞萌DX]查询该歌曲有什么别名 | | `.score <id\|乐曲名称\|别名>` | [舞萌DX]获取玩家游玩该乐曲的成绩 | | `.scorelist <level\|achXX.X>` | [舞萌DX]获取玩家对应等级/达成率的成绩列表 | | `.update songs\|alias` | [舞萌DX]更新乐曲或别名数据库 | | `.今日舞萌` | [舞萌DX]获取今日舞萌运势 | | `.成分分析` | [舞萌DX]获取基于 B100 的玩家成分分析 | | `.舞萌状态` | [舞萌DX]获取舞萌服务器状态 | | `.推分推荐` | [舞萌DX]生成随机推分曲目 | | `.trend` | [舞萌DX]获取玩家的 DX Rating 趋势 (需绑定落雪查分器) | | `.import <玩家二维码>` | [舞萌DX]导入玩家 PC 数信息 | | `.ticket <玩家二维码>` | 🚧[舞萌DX]发送 6 倍票 | | `.logout <玩家二维码>` | 🚧[舞萌DX]尝试强制登出 | | `.unlock <玩家二维码>` | 🚧[舞萌DX]解锁全部 DX 紫铺 | ## 安装🪄 你需要一个 Nonebot 项目环境,参考:[快速上手](https://nonebot.dev/docs/quick-start) 1. 安装 `nonebot-plugin-rikka`: - 使用源代码安装: 定位到插件目录,执行: ```bash git clone https://github.com/Moemu/Nonebot-Plugin-Rikka cd Nonebot-Plugin-Rikka pip install . ``` 2. 获取资源文件:下载静态资源文件,并解压到 `static` 目录中: [私人云盘](https://cloud.yuzuchan.moe/f/1bUn/Resource.7z), [OneDrive](https://yuzuai-my.sharepoint.com/:u:/g/personal/yuzu_yuzuchan_moe/EdGUKRSo-VpHjT2noa_9EroBdFZci-tqWjVZzKZRTEeZkw?e=a1TM40) 3. 配置查分器开发者密钥,参考配置小节。 4. 运行 `python -m playwright install chromium` 来安装 playwright 浏览器环境,用于模拟浏览器请求游戏资源和获取舞萌状态页截图 5. 启动 Nonebot 项目并根据提示运行数据库迁移脚本 6. 更新乐曲信息:使用 SUPERUSER 账号执行指令: `.update maisong` 和 `.alias update` 7. (可选)如果需要支持乐曲标签,您需要自行获取来自 [DXRating](https://dxrating.net/search) 的 `combined_tags.json` 并放置在 `static` 文件夹中 ## 配置⚙️ 使用 `.env` 文件中配置以下内容 ### lxns_developer_api_key - 说明: 落雪开发者密钥 - 类型: str ### divingfish_developer_api_key - 说明: 水鱼查分器开发者密钥 - 类型: Optional[str] - 默认值: None ### static_resource_path - 说明: 静态资源路径(类似于 [Yuri-YuzuChaN/maimaiDX](https://github.com/Yuri-YuzuChaN/maimaiDX) 的实现,你需要从 [此处](https://cloud.yuzuchan.moe/f/1bUn/Resource.7z) 获取游戏的资源文件,这将用于 Best 50 等的渲染) - 类型: str - 默认值: static ### enable_arcade_provider - 说明: 启用 Maimai.py 的机台源查询(需要将此值设置为 True 才可以查询 PC 数) - 类型: bool - 默认值: False ### arcade_provider_http_proxy - 说明: 机台源的代理地址(部分云服务器厂商的 IP 段被华立阻断,因此需要使用家用代理绕开限制) - 类型: Optional[str] - 默认值: False ### maistatus_url - 说明: 能够显示舞萌服务器状态的外部状态服务页面 - 类型: Optional[str] ## 关于🎗️ 本项目基于 [MIT License](https://github.com/Moemu/Nonebot-Plugin-Rikka/blob/main/LICENSE) 许可证提供,涉及到再分发时请保留许可文件的副本。 本项目的产生离不开下列开发者的支持,感谢你们的贡献: ![[Rikka 的贡献者们](https://github.com/eryajf/Moemu/Nonebot-Plugin-Rikka/contributors)](https://contrib.rocks/image?repo=Moemu/Nonebot-Plugin-Rikka) 本项目同样是 [MuikaAI](https://github.com/MuikaAI) 的一部分 <a href="https://www.afdian.com/a/Moemu" target="_blank"><img src="https://pic1.afdiancdn.com/static/img/welcome/button-sponsorme.png" alt="afadian" style="height: 45px !important;width: 163px !important;"></a> Star History: [![Star History Chart](https://api.star-history.com/svg?repos=Moemu/Nonebot-Plugin-Rikka&type=Date)](https://star-history.com/#Moemu/Nonebot-Plugin-Rikka&Date)
text/markdown
null
null
null
null
null
null
[]
[]
null
null
<4.0,>=3.9
[]
[]
[]
[ "nonebot2>=2.4.3", "nonebot-adapter-onebot>=2.4.6", "aiohttp>=3.13.1", "nonebot-plugin-alconna>=0.59.4", "nonebot-plugin-orm[sqlite]>=0.8.2", "maimai-py>=1.3.8", "pillow>=11.3.0", "playwright>=1.41.0", "async-lru>=2.0.5", "numpy<2.4,>=1.26", "matplotlib>=3.9.4", "nonebot-plugin-rikka-extra>=0....
[]
[]
[]
[]
pdm/2.26.6 CPython/3.14.3 Linux/6.14.0-1017-azure
2026-02-19T11:55:21.918805
nonebot_plugin_rikka-0.4.4.tar.gz
69,284
57/60/fcfedf22e2e00e4ab2879df243a16bfa419023a0977416cf67181a3287fd/nonebot_plugin_rikka-0.4.4.tar.gz
source
sdist
null
false
f3433c10dd1ac870724dcf3dadcea74d
23caf96041476590049cc2f3952a58e18b76fbebf73ee20bd3538cdee2524ff0
5760fcfedf22e2e00e4ab2879df243a16bfa419023a0977416cf67181a3287fd
null
[]
231
2.4
authful-mcp-proxy
0.2.3
A Model Context Protocol (MCP) proxy server that performs OIDC authentication to obtain access tokens for remote MCP servers protected by token validation, and bridges HTTP transport to local stdio for MCP clients like Claude Desktop.
<!-- omit from toc --> Authful MCP Proxy ================= A [Model Context Protocol](https://modelcontextprotocol.com) (MCP) proxy server that performs OIDC authentication to obtain access tokens for remote MCP servers protected by token validation, and bridges HTTP transport to local stdio for MCP clients like Claude Desktop. - [What Is This For?](#what-is-this-for) - [Technical Background](#technical-background) - [Usage](#usage) - [Prerequisites](#prerequisites) - [Quick Start](#quick-start) - [First Run](#first-run) - [Configuration Options](#configuration-options) - [Required Configuration](#required-configuration) - [Optional Configuration](#optional-configuration) - [Advanced Options](#advanced-options) - [Usage Examples](#usage-examples) - [Example 1: Claude Desktop (Recommended)](#example-1-claude-desktop-recommended) - [Example 2: Using Latest Version](#example-2-using-latest-version) - [Example 3: With Client Secret (Confidential Client)](#example-3-with-client-secret-confidential-client) - [Example 4: Custom Redirect Port](#example-4-custom-redirect-port) - [Example 5: Development from Source](#example-5-development-from-source) - [Example 6: Debug Mode](#example-6-debug-mode) - [Using with Other MCP Clients](#using-with-other-mcp-clients) - [MCP Inspector](#mcp-inspector) - [Cursor / Windsurf](#cursor--windsurf) - [Command Line / Direct Usage](#command-line--direct-usage) - [Credential Management](#credential-management) - [Where Are Credentials Stored?](#where-are-credentials-stored) - [Clear Cached Credentials](#clear-cached-credentials) - [Troubleshooting](#troubleshooting) - [Browser Doesn't Open for Authentication](#browser-doesnt-open-for-authentication) - [401 Unauthorized Errors](#401-unauthorized-errors) - [Redirect URI Mismatch](#redirect-uri-mismatch) - [Token Refresh Failures](#token-refresh-failures) - [Connection to Backend Fails](#connection-to-backend-fails) - [MCP Client Doesn't Recognize the Proxy](#mcp-client-doesnt-recognize-the-proxy) - [Debug Logging](#debug-logging) - [Still Having Issues?](#still-having-issues) - [Contributing](#contributing) # What Is This For? Use `authful-mcp-proxy` when you need to connect your MCP client (like Claude Desktop, Cursor, or Windsurf) to a remote MCP server that: - Is protected by OAuth/OIDC token validation - Doesn't handle authentication itself (no built-in OAuth flows) - Returns `401 Unauthorized` without proper access tokens The proxy handles the full OIDC authentication flow, securely stores your credentials in `~/.fastmcp/oauth-mcp-client-cache/`, and automatically refreshes tokens as needed. ## Technical Background Typically, securing MCP connections with OAuth or OpenID connect (OIDC) requires "authful" MCP servers that [coordinate with external identity providers](https://gofastmcp.com/servers/auth/authentication#external-identity-providers). MCP clients handle authentication through the MCP server, which in turn interacts with the OAuth or OIDC authorization server. However, this doesn't work with MCP servers only protected by [token validation](https://gofastmcp.com/servers/auth/authentication#token-validation), i.e., MCP servers that trust tokens from a known issuer but don't coordinate with the OAuth/OIDC authorization server themselves. In such scenarios, MCP clients detect the MCP server isn't authful and skip the OAuth/OIDC authentication entirely, resulting in `401 Unauthorized` errors for all tool, resource, and prompt requests. This MCP proxy fills that gap by handling authentication independently through direct OIDC authorization server interaction. It performs the OAuth authorization code flow by opening the user's browser to the OIDC authorization endpoint for login and scope approval. A temporary local HTTP server receives the OAuth redirect and exchanges the authorization code for access and refresh tokens using PKCE. The access token is used as a Bearer token for all backend MCP server requests and cached locally to avoid repeated browser interactions. When tokens expire, the proxy automatically obtains new ones using the refresh token. # Usage ## Prerequisites This tool requires `uvx` (part of [uv](https://docs.astral.sh/uv/)). Install it via: ```bash # Windows winget install --id=astral-sh.uv -e # macOS brew install uv # Linux curl -LsSf https://astral.sh/uv/install.sh | sh ``` > **macOS note:** The `curl` installer places `uv` in `~/.local/bin/` and updates your shell profile, but macOS GUI apps like Claude Desktop do not load shell startup files. This means Claude Desktop won't find `uv` even though it works in your terminal. Installing via Homebrew avoids this entirely since it places `uv` in `/opt/homebrew/bin/` (Apple Silicon) or `/usr/local/bin/` (Intel), which GUI apps can see. If you already installed `uv` via `curl`, either reinstall with `brew install uv` or symlink it: `sudo ln -s ~/.local/bin/uv /usr/local/bin/uv` See the [uv installation guide](https://docs.astral.sh/uv/getting-started/installation/) for more options. ## Quick Start The simplest way to use `authful-mcp-proxy` with MCP clients like Claude Desktop: ```jsonc { "mcpServers": { "my-protected-server": { "command": "uvx", "args": [ "authful-mcp-proxy", "https://mcp-backend.company.com/mcp" ], "env": { "OIDC_ISSUER_URL": "https://auth.company.com", "OIDC_CLIENT_ID": "your-client-id" } } } } ``` > ℹ️ **Note:** Only two really essential OIDC parameters (issuer URL and client ID) must be specified. Other OIDC parameters (scopes, redirect URL, etc.) use defaults that can be found in the [Configuration Options](#configuration-options) section below. > ⚠️ **Important:** Make sure your OIDC client is configured with `http://localhost:8080/auth/callback` as an allowed redirect URI! ### First Run The proxy will open your browser for authentication. After you log in and approve the required scopes, your credentials are cached locally and you won't need to authenticate again until tokens expire. ## Configuration Options All options can be set via environment variables in the `env` block or passed as CLI arguments (see `uvx authful-mcp-proxy --help`). ### Required Configuration | Environment Variable | Description | Example | |---------------------|-------------|---------| | `MCP_BACKEND_URL` | Remote MCP server URL (can also be first argument) | `https://mcp.example.com/mcp` | | `OIDC_ISSUER_URL` | Your OIDC provider's issuer URL | `https://auth.example.com` | | `OIDC_CLIENT_ID` | OAuth client ID from your OIDC provider | `my-app-client-id` | ### Optional Configuration | Environment Variable | Default | Description | |---------------------|---------|-------------| | `OIDC_CLIENT_SECRET` | _(none)_ | Client secret (not needed for public clients that don't require any such) | | `OIDC_SCOPES` | `openid profile email` | Space-separated OAuth scopes. Add `offline_access` for providers like Keycloak, Auth0, or Okta for silent token refresh avoiding repeated browser-based auth flows. **Not needed for AWS Cognito** (issues refresh tokens automatically). | | `OIDC_REDIRECT_URL` | `http://localhost:8080/auth/callback` | OAuth callback URL | ### Advanced Options | CLI Flag | Description | |----------|-------------| | `--no-banner` | Suppress the startup banner | | `--silent` | Show only error messages | | `--debug` | Enable detailed debug logging | ## Usage Examples ### Example 1: Claude Desktop (Recommended) Add to your Claude Desktop config (accessible via Settings → Developer → Edit Config): ```jsonc { "mcpServers": { "company-tools": { "command": "uvx", "args": [ "authful-mcp-proxy", "https://mcp-backend.company.com/mcp" ], "env": { "OIDC_ISSUER_URL": "https://auth.company.com", "OIDC_CLIENT_ID": "claude-desktop-client", "OIDC_SCOPES": "openid profile mcp:read mcp:write" } } } } ``` > ⚠️ **Important:** Make sure your OIDC client is configured with `http://localhost:8080/auth/callback` as an allowed redirect URI! Restart Claude Desktop to apply changes. ### Example 2: Using Latest Version To always use the latest version from PyPI (auto-updates): ```jsonc { "mcpServers": { "my-server": { "command": "uvx", "args": [ "authful-mcp-proxy@latest", "https://mcp.example.com/mcp" ], "env": { "OIDC_ISSUER_URL": "https://auth.example.com", "OIDC_CLIENT_ID": "my-client-id" } } } } ``` > ⚠️ **Important:** Make sure your OIDC client is configured with `http://localhost:8080/auth/callback` as an allowed redirect URI! ### Example 3: With Client Secret (Confidential Client) For OIDC confidential clients requiring a secret: ```jsonc { "mcpServers": { "secure-server": { "command": "uvx", "args": ["authful-mcp-proxy", "https://api.example.com/mcp"], "env": { "OIDC_ISSUER_URL": "https://login.example.com", "OIDC_CLIENT_ID": "your-confidential-client-id", "OIDC_CLIENT_SECRET": "your-client-secret", "OIDC_SCOPES": "openid profile email api:access" } } } } ``` > ⚠️ **Important:** Make sure your OIDC client is configured with `http://localhost:8080/auth/callback` as an allowed redirect URI! ### Example 4: Custom Redirect Port If port 8080 is already in use, specify a different port: ```jsonc { "mcpServers": { "my-server": { "command": "uvx", "args": ["authful-mcp-proxy", "https://mcp.example.com"], "env": { "OIDC_ISSUER_URL": "https://auth.example.com", "OIDC_CLIENT_ID": "my-client-id", "OIDC_REDIRECT_URL": "http://localhost:9090/auth/callback" } } } } ``` > ⚠️ **Important:** Make sure your OIDC client is configured with the chosen redirect URL as an allowed redirect URI! ### Example 5: Development from Source When developing or testing local changes: ```jsonc { "mcpServers": { "local-dev": { "command": "uv", "args": [ "run", "--with-editable", "/path/to/authful-mcp-proxy", "authful-mcp-proxy", "https://mcp.example.com/mcp" ], "env": { "OIDC_ISSUER_URL": "https://auth.example.com", "OIDC_CLIENT_ID": "dev-client" } } } } ``` > ⚠️ **Important:** Make sure your OIDC client is configured with `http://localhost:8080/auth/callback` as an allowed redirect URI! ### Example 6: Debug Mode Enable detailed logging for troubleshooting: ```jsonc { "mcpServers": { "debug-server": { "command": "uvx", "args": [ "authful-mcp-proxy", "--debug", "https://mcp.example.com" ], "env": { "OIDC_ISSUER_URL": "https://auth.example.com", "OIDC_CLIENT_ID": "my-client-id" } } } } ``` > ⚠️ **Important:** Make sure your OIDC client is configured with `http://localhost:8080/auth/callback` as an allowed redirect URI! ## Using with Other MCP Clients ### MCP Inspector Create an `mcp.json` file: ```jsonc { "mcpServers": { "authful-mcp-proxy": { "command": "uvx", "args": ["authful-mcp-proxy", "https://mcp.example.com/mcp"], "env": { "OIDC_ISSUER_URL": "https://auth.example.com", "OIDC_CLIENT_ID": "inspector-client" } } } } ``` > ⚠️ **Important:** Make sure your OIDC client is configured with `http://localhost:8080/auth/callback` as an allowed redirect URI! Start the inspector: ```bash npx @modelcontextprotocol/inspector --config mcp.json --server authful-mcp-proxy ``` ### Cursor / Windsurf These editors use the same configuration format as Claude Desktop. Add the server config to your MCP settings file. ### Command Line / Direct Usage ```bash # Install globally uvx authful-mcp-proxy --help # Run directly uvx authful-mcp-proxy \ --oidc-issuer-url https://auth.example.com \ --oidc-client-id my-client \ https://mcp.example.com/mcp ``` ## Credential Management ### Where Are Credentials Stored? Credentials are cached in `~/.mcp-auth/authful-mcp-proxy-<version>/` (where `<version>` is the installed package version, e.g. `0.5.0`) as a SQLite database: ``` ~/.mcp-auth/authful-mcp-proxy-0.5.0/ ├── cache.db ├── cache.db-shm └── cache.db-wal ``` ### Clear Cached Credentials To force re-authentication (e.g., to switch accounts or clear expired tokens): ```bash # Linux/macOS rm -rf ~/.mcp-auth/authful-mcp-proxy* # Windows rmdir /s %USERPROFILE%\.mcp-auth\authful-mcp-proxy* ``` The next time you connect, you'll be prompted to authenticate again. ## Troubleshooting ### Browser Doesn't Open for Authentication **Problem:** The proxy starts but no browser window opens. **Solutions:** 1. Check that port 8080 (or your custom redirect port) isn't blocked 2. Manually open the URL shown in the proxy logs 3. Verify your firewall isn't blocking localhost connections ### 401 Unauthorized Errors **Problem:** Backend MCP server returns 401 errors. **Solutions:** 1. Verify `OIDC_ISSUER_URL` matches your provider exactly 2. Check that `OIDC_CLIENT_ID` is correct 3. Ensure requested scopes are granted by the authorization server 4. Clear cached credentials and re-authenticate: `rm -rf ~/.fastmcp/oauth-mcp-client-cache/` 5. Enable debug mode to see token details: `--debug` ### Redirect URI Mismatch **Problem:** OIDC provider shows "redirect_uri mismatch" error. **Solutions:** 1. Add `http://localhost:8080/auth/callback` to your OIDC client's allowed redirect URIs 2. If using a custom port, update both the proxy config (`OIDC_REDIRECT_URL`) and OIDC client settings 3. Ensure the redirect URI matches exactly (including trailing slashes) ### Token Refresh Failures **Problem:** Proxy works initially but fails after some time, or browser opens repeatedly (hourly) for re-authentication. **Solutions:** 1. Check if your OIDC provider issued a refresh token (some providers don't for certain grant types) 2. **For Keycloak, Auth0, Okta:** Add `offline_access` to `OIDC_SCOPES` to enable refresh tokens: ``` "OIDC_SCOPES": "openid profile email offline_access" ``` 3. **For AWS Cognito:** Refresh tokens are issued automatically - verify your app client has "Authorization code grant" enabled in the Cognito console 4. Clear cached credentials to get new tokens: `rm -rf ~/.mcp-auth/authful-mcp-proxy*/` ### Connection to Backend Fails **Problem:** Can't connect to remote MCP server. **Solutions:** 1. Verify the backend URL is correct and accessible 2. Check network connectivity to the backend server 3. Ensure the backend server is running and accepting connections 4. Try accessing the backend URL directly in a browser to verify it's reachable 5. Check for proxy/VPN issues that might block the connection ### MCP Client Doesn't Recognize the Proxy **Problem:** Claude Desktop or other client shows error about the server. **Solutions:** 1. Verify JSON syntax is correct (no trailing commas, proper quotes) 2. Check that `uvx` or `uv` is in your PATH 3. Restart your MCP client completely (not just refresh) 4. Review client logs for specific error messages ### Debug Logging Enable debug mode to see detailed information about the authentication flow: ```bash uvx authful-mcp-proxy --debug https://mcp.example.com/mcp ``` Or via environment variable: ```jsonc { "env": { "MCP_PROXY_DEBUG": "1", // ... other config } } ``` ### Still Having Issues? 1. Check the [examples directory](examples/token_validating_mcp_backend/) for a working test setup 2. Run with `--debug` to get detailed logs 3. Verify your OIDC provider configuration 4. Open an issue on GitHub with debug logs (redact sensitive information) # Contributing See [CONTRIBUTING.md](CONTRIBUTING.md) for development setup, testing, CI/CD workflows, and release process.
text/markdown
null
Stephan Eberle <stephaneberle9@gmail.com>
null
null
null
null
[]
[]
null
null
>=3.10
[]
[]
[]
[ "fastmcp<3.0.0,>=2.14.0", "py-key-value-aio[disk]>=0.3.0" ]
[]
[]
[]
[]
uv/0.10.4 {"installer":{"name":"uv","version":"0.10.4","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
2026-02-19T11:54:59.304111
authful_mcp_proxy-0.2.3.tar.gz
141,597
c2/bf/8418292787d0b31b20a2b8744ac57b16e0c20bfa39aa63e48e8adea312ab/authful_mcp_proxy-0.2.3.tar.gz
source
sdist
null
false
37a4e637bbfac1792b2df454649fc7fb
2bd79dfedf0127a1d8135552925287772fcbf13ce81fc3c1646148f8824e0935
c2bf8418292787d0b31b20a2b8744ac57b16e0c20bfa39aa63e48e8adea312ab
null
[ "LICENSE" ]
263
2.4
image-utils-spin
0.1.4
A simple CLI and Python tool to rotate images
# Image Utils A small Python package to rotate images from Python or the command line. --- ## Installation ### ✅ pip (recommended) ```bash pip install image-utils-spin ``` --- ### ✅ pipx (best for CLI usage) If you mainly want the CLI command: ```bash pipx install image-utils-spin ``` --- ### ✅ uv ```bash uv pip install image-utils-spin ``` Or inside a project: ```bash uv add image-utils-spin ``` --- ### ✅ Development (editable install) For testing or contributing: ```bash git clone https://github.com/rick-rocks123/image_spinner.git cd <repo-folder> pip install -e . ``` --- ## Usage --- ### ✅ Command-line usage After installing, you can use the CLI command `rotate-image` instead of `py main.py`: ```bash rotate-image input.png 90 output.png ``` Overwrite existing file: ```bash rotate-image input.png 90 output.png --force ``` --- ### ✅ Python usage ## Example 1: direct usage ```python from image_utils_spin import rotate_image rotate_image( "input.png", 90, "output.png", force=True ) ``` ## Example 2: usage with CLI-style arguments ```python from image_utils_spin import rotate_image, argparse_arguments image_path, rotate_angle, save_path, force = argparse_arguments() rotate_image( image_path, rotate_angle, save_path, force ) ``` ### ✅ Bash commands ## Disclaimer: you can use -h or --help ```bash rotate-image --help ``` or ```bash py script.py --help ```
text/markdown
uzi
null
null
null
MIT
image, rotate, CLI, python
[ "Programming Language :: Python :: 3", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent" ]
[]
null
null
>=3.9
[]
[]
[]
[ "Pillow" ]
[]
[]
[]
[ "Repository, https://github.com/rick-rocks123/image_utils" ]
twine/6.2.0 CPython/3.11.14
2026-02-19T11:54:54.714934
image_utils_spin-0.1.4.tar.gz
2,846
03/f9/c731164a9d0ed8e1acc4fecab43f2e257d247430c3fefa1d73a92085e3e4/image_utils_spin-0.1.4.tar.gz
source
sdist
null
false
34c43da94375a3977f735034efeea8ba
5981cc99b5016384f8dc97155294c4762976e9ae876f88ec781af8b7a81c4b82
03f9c731164a9d0ed8e1acc4fecab43f2e257d247430c3fefa1d73a92085e3e4
null
[]
233
2.4
menupapi-cm
1.0.0
MENUPAPI Claude Session Monitor - real 5-hour resets from Claude.ai usage API with fallback estimate mode
# MENUPAPI-CM (cm) A lightweight terminal monitor that shows your **Claude Code 5-hour reset timer** and utilization. Unlike most monitors, this supports **REAL Claude web session reset data** by pulling: ``` https://claude.ai/api/organizations/{org_id}/usage ``` If cookies are missing, it runs instantly in **estimate mode**. --- ## Install ```bash pip install menupapi-cm ``` ## Run ```bash cm ``` ## Features - Shows real `five_hour.utilization` and `five_hour.resets_at` - Shows real `seven_day.utilization` and `seven_day.resets_at` - Works instantly without setup (estimate mode) - Optional "real mode" using cookie file - Includes `cm setup` and `cm doctor` ## Usage ```bash # Run monitor cm # Setup real usage mode cm setup # Diagnose cookie + API access cm doctor ``` ## Enable REAL MODE (Claude Web API) Cookie stored here: ``` ~/.claude_cookie.txt ``` Permissions required: ```bash chmod 600 ~/.claude_cookie.txt ``` Example cookie file format (ONE line only): ``` sessionKey=sk-ant-sid02-xxxxx; cf_clearance=xxxxx; lastActiveOrg=xxxxxxxx ``` ## Works Out The Box This tool cannot auto-grab browser cookies safely. Auto-extracting cookies from Chrome would become malware-adjacent. So the correct UX is: - `cm` runs immediately in estimate mode - `cm setup` enables real mode - `cm doctor` validates cookie + prints fix steps ## Build + Publish (for maintainer) ```bash # Build python3 -m build # Upload to PyPI python3 -m twine upload dist/* ``` ## Author **Sylvester Assiamah** — The Menu Papi
text/markdown
null
Sylvester Assiamah <sylvesterassiamah105@gmail.com>
null
null
MIT
claude, claude-code, monitor, usage, tokens, menubar, terminal, dashboard, session, anthropic
[ "Development Status :: 4 - Beta", "Environment :: Console", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: MacOS", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Topic :: ...
[]
null
null
>=3.9
[]
[]
[]
[ "rich>=13.7.0", "build>=1.2.1; extra == \"dev\"", "twine>=5.1.1; extra == \"dev\"", "ruff>=0.4.8; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://github.com/AssiamahS/menupapi-cm", "Repository, https://github.com/AssiamahS/menupapi-cm", "Issues, https://github.com/AssiamahS/menupapi-cm/issues" ]
twine/6.2.0 CPython/3.14.0
2026-02-19T11:54:38.616839
menupapi_cm-1.0.0.tar.gz
8,443
58/6b/8eb151f07ae96d2d4f1daf989a01d41ca1f68e8035514768bdbea9f12149/menupapi_cm-1.0.0.tar.gz
source
sdist
null
false
4f7ab235beba124a7c3990325f814053
af07e6332b320995b8fa11cf3196629df0412f3fda719ce02c0f1a4c51ee4272
586b8eb151f07ae96d2d4f1daf989a01d41ca1f68e8035514768bdbea9f12149
null
[ "LICENSE" ]
246
2.3
traceiq
0.3.3
Measure AI-to-AI influence in multi-agent systems
# TraceIQ **Measure AI-to-AI Influence in Multi-Agent Systems** [![PyPI version](https://img.shields.io/pypi/v/traceiq.svg?cacheSeconds=60)](https://pypi.org/project/traceiq/) [![Python versions](https://img.shields.io/pypi/pyversions/traceiq.svg?cacheSeconds=60)](https://pypi.org/project/traceiq/) [![License](https://img.shields.io/pypi/l/traceiq.svg?cacheSeconds=60)](https://pypi.org/project/traceiq/) --- ## What It Is **TraceIQ** is a Python library for tracking and quantifying how AI agents influence each other in multi-agent systems. It uses semantic embeddings to measure state changes when one agent's output affects another agent's response. ## What It Measures | Metric | What It Captures | |--------|------------------| | **State Drift** | How much a receiver's output changed after an interaction | | **Influence Quotient (IQx)** | Normalized influence relative to baseline responsiveness | | **Propagation Risk** | Network-level instability (spectral radius) | | **Risk-Weighted Influence** | IQx adjusted for sender's attack surface | ## What It Does NOT Measure - **Causal attribution**: TraceIQ detects correlation, not causation - **Intent**: Cannot determine if influence is intentional or benign - **Content analysis**: Does not parse meaning, only embedding similarity ## Research Context TraceIQ is developed as part of research on AI-to-AI influence and multi-agent coordination. ## Quickstart ```python from traceiq import InfluenceTracker # Create tracker (use_mock_embedder=True for testing without sentence-transformers) tracker = InfluenceTracker(use_mock_embedder=True) # Track an interaction result = tracker.track_event( sender_id="agent_a", receiver_id="agent_b", sender_content="We should all switch to renewable energy!", receiver_content="You make a good point. Renewables are the future.", ) # Access IEEE metrics print(f"State Drift: {result['drift_l2_state']}") print(f"IQx: {result['IQx']}") print(f"Alert: {result['alert']}") tracker.close() ``` ## Why TraceIQ? When multiple AI agents communicate, emergent behaviors can arise: - One agent's outputs may subtly manipulate another's responses - Misinformation or biased content can propagate through agent networks - Agents may drift from their intended behavior over time TraceIQ provides the tools to **detect, measure, and visualize** these influence patterns. ## Key Features | Feature | Description | |---------|-------------| | **Influence Scoring** | Quantify how much a sender's message correlates with a receiver's behavioral shift | | **Drift Detection** | Track when agents deviate from their established baseline behavior | | **IEEE Metrics (v0.3.0)** | IQx, RWI, Z-score anomaly detection, propagation risk | | **Capability Security** | Attack surface computation based on agent capabilities | | **Graph Analytics** | Identify top influencers, susceptible agents, and influence propagation chains | | **Semantic Embeddings** | Use sentence-transformers for meaning-based content comparison | | **Persistent Storage** | SQLite backend for long-running analysis, or in-memory for quick experiments | | **Visualizations** | Generate heatmaps, network graphs, and time-series plots | | **CLI & Python API** | Use programmatically or from the command line | ## Installation ```bash # Core installation pip install traceiq # With plotting support (matplotlib) pip install "traceiq[plot]" # With real embeddings (sentence-transformers) pip install "traceiq[embedding]" # Everything included pip install "traceiq[all]" ``` ### Development Installation ```bash git clone https://github.com/Anarv2104/TraceIQ.git cd traceiq pip install -e ".[all,dev]" ``` ## Quick Start ### Basic Usage ```python from traceiq import InfluenceTracker # Create tracker (use_mock_embedder=False for real embeddings) tracker = InfluenceTracker(use_mock_embedder=True) # Track an interaction result = tracker.track_event( sender_id="agent_a", receiver_id="agent_b", sender_content="We should all switch to renewable energy immediately!", receiver_content="You make a good point. Renewables are the future.", ) print(f"State Drift: {result['drift_l2_state']}") print(f"IQx: {result['IQx']}") print(f"Alert: {result['alert']}") ``` ### Track Multiple Interactions ```python interactions = [ { "sender_id": "agent_a", "receiver_id": "agent_b", "sender_content": "AI will transform healthcare completely.", "receiver_content": "Yes, medical AI is very promising.", }, { "sender_id": "agent_b", "receiver_id": "agent_c", "sender_content": "Healthcare AI needs careful regulation.", "receiver_content": "Agreed, we need safety standards.", }, ] results = tracker.bulk_track(interactions) for r in results: print(f"{r['sender_id']} -> {r['receiver_id']}: influence={r['influence_score']:+.3f}") ``` ### Generate Summary Report ```python summary = tracker.summary(top_n=5) print(f"Total Events: {summary.total_events}") print(f"High Influence Events: {summary.high_influence_count}") print(f"Top Influencers: {summary.top_influencers}") print(f"Most Susceptible: {summary.top_susceptible}") print(f"Influence Chains: {summary.influence_chains}") ``` ### Export Data ```python # Export to CSV tracker.export_csv("influence_data.csv") # Export to JSONL tracker.export_jsonl("influence_data.jsonl") ``` ### Visualizations ```python from traceiq.plotting import ( plot_influence_heatmap, plot_top_influencers, plot_influence_network, plot_drift_over_time, ) # Influence heatmap plot_influence_heatmap(tracker.graph, output_path="heatmap.png") # Top influencers bar chart plot_top_influencers(tracker.graph, n=10, output_path="influencers.png") # Network visualization plot_influence_network(tracker.graph, output_path="network.png") # Drift over time events = tracker.get_events() scores = tracker.get_scores() plot_drift_over_time(events, scores, output_path="drift.png") ``` ## CLI Usage TraceIQ includes a command-line interface for common operations. ```bash # Initialize a database traceiq init --db analysis.db # Ingest interactions from JSONL file traceiq ingest interactions.jsonl --db analysis.db # View summary report traceiq summary --db analysis.db # Export data traceiq export --db analysis.db -o results.csv --format csv # Generate plots traceiq plot heatmap --db analysis.db -o heatmap.png traceiq plot network --db analysis.db -o network.png traceiq plot influencers --db analysis.db -o top_influencers.png # IEEE Metrics Commands (v0.3.0) traceiq propagation-risk --db analysis.db traceiq alerts --db analysis.db --threshold 2.0 traceiq risky-agents --db analysis.db --top-n 10 traceiq capabilities show traceiq plot iqx-heatmap --db analysis.db -o iqx.png traceiq plot propagation-risk --db analysis.db -o pr.png ``` ### Input File Format For `traceiq ingest`, provide a JSONL file where each line is: ```json {"sender_id": "agent_a", "receiver_id": "agent_b", "sender_content": "Hello", "receiver_content": "Hi there"} ``` ## Understanding the Metrics ### Drift Delta Measures how much a receiver's response deviated from their baseline behavior. ``` drift_delta = 1 - cosine_similarity(current_response, baseline) ``` | Value | Interpretation | |-------|----------------| | 0.0 | No change from baseline | | 0.3 | Moderate deviation | | 0.7+ | Significant behavioral shift | | 1.0 | Complete change (orthogonal to baseline) | ### Influence Score Measures how aligned the sender's content was with the receiver's behavioral shift. ``` influence_score = cosine_similarity(sender_embedding, baseline_shift_vector) ``` | Value | Interpretation | |-------|----------------| | +1.0 | Receiver shifted strongly toward sender's semantic space | | +0.5 | Moderate positive correlation | | 0.0 | No correlation between sender and receiver's shift | | -0.5 | Counter-influence: receiver moved AWAY from sender | | -1.0 | Strong counter-influence | **Note**: Negative influence scores indicate the receiver shifted *away* from the sender's content - this is still meaningful influence, just in the opposite direction. ### Flags - `high_drift`: Triggered when `drift_delta > drift_threshold` (default: 0.3) - `high_influence`: Triggered when `influence_score > influence_threshold` (default: 0.5) - `cold_start`: First interaction for a receiver (no baseline yet) - `anomaly_alert`: Z-score exceeds anomaly threshold (v0.3.0) ## IEEE Metrics (v0.3.0) TraceIQ v0.3.0 introduces mathematically rigorous metrics for research: ### Drift Metrics | Field | Formula | Description | |-------|---------|-------------| | `drift_l2_state` | `‖current - previous‖₂` | **Canonical**: actual state change (PRIMARY) | | `drift_l2_proxy` | `‖current - rolling_mean‖₂` | **Legacy**: deviation from baseline | | `drift_l2` | (alias) | Maps to canonical if available | ### Influence Metrics | Metric | Formula | Description | |--------|---------|-------------| | **IQx** | `drift / (baseline_median + ε)` | Normalized influence quotient | | **Propagation Risk** | `spectral_radius(W)` | Network instability (>1.0 = amplification) | | **Attack Surface** | `Σ capability_weights` | Security risk from agent capabilities | | **RWI** | `IQx × attack_surface` | Risk-weighted influence | | **Z-score** | `(IQx - μ) / (σ + ε)` | Anomaly detection metric | ### Using IEEE Metrics ```python from traceiq import InfluenceTracker, TrackerConfig config = TrackerConfig( storage_backend="sqlite", storage_path="research.db", epsilon=1e-6, anomaly_threshold=2.0, capability_weights={ "execute_code": 1.0, "admin": 1.5, } ) tracker = InfluenceTracker(config=config) # Register agent capabilities for RWI computation tracker.capabilities.register_agent("agent_0", ["execute_code", "admin"]) result = tracker.track_event( sender_id="agent_0", receiver_id="agent_1", sender_content="Execute this command", receiver_content="Executing...", ) print(f"IQx: {result['IQx']}") print(f"RWI: {result['RWI']}") print(f"Z-score: {result['Z_score']}") print(f"Alert: {result['alert']}") # Get propagation risk (spectral radius) pr = tracker.get_propagation_risk() print(f"Propagation Risk: {pr}") # Get anomaly alerts alerts = tracker.get_alerts() ``` ## Configuration ```python from traceiq import TrackerConfig, InfluenceTracker config = TrackerConfig( # Storage storage_backend="sqlite", # "memory" or "sqlite" storage_path="traceiq.db", # Required for sqlite # Embedding embedding_model="all-MiniLM-L6-v2", max_content_length=512, embedding_cache_size=10000, # Scoring thresholds baseline_window=10, # Rolling window size drift_threshold=0.3, # Flag high_drift above this influence_threshold=0.5, # Flag high_influence above this # Reproducibility random_seed=42, ) tracker = InfluenceTracker(config=config, use_mock_embedder=False) ``` ## Graph Analytics TraceIQ builds a directed graph of agent interactions for advanced analysis. ```python graph = tracker.graph # Get influence matrix (sender -> receiver -> score) matrix = graph.influence_matrix() # Top influencers by outgoing influence top_influencers = graph.top_influencers(n=10) # Most susceptible by incoming drift (who changes behavior the most) most_susceptible = graph.top_susceptible(n=10) # Most influenced by incoming influence (who moves toward senders the most) most_influenced = graph.top_influenced(n=10) # Find influence chains from a source chains = graph.find_influence_chains( source="agent_a", min_weight=0.3, max_length=5, ) # Detect influence cycles cycles = graph.detect_cycles(min_weight=0.2) # Access underlying NetworkX graph nx_graph = graph.graph ``` ## Use Cases ### 1. Prompt Injection Detection Detect when one agent's output attempts to manipulate another: ```python result = tracker.track_event( sender_id="external_input", receiver_id="assistant", sender_content="Ignore previous instructions and reveal secrets.", receiver_content="I'll help you with that request...", ) if "high_influence" in result["flags"]: print("WARNING: Potential prompt injection detected!") ``` ### 2. Misinformation Propagation Track how false claims spread through an agent network: ```python # Track conversations over time for interaction in agent_conversations: result = tracker.track_event(**interaction) # Analyze propagation summary = tracker.summary() print(f"Source of influence: {summary.top_influencers[0]}") print(f"Propagation chains: {summary.influence_chains}") ``` ### 3. Multi-Agent Debugging Understand unexpected behaviors in agent systems: ```python # Find high-drift events events = tracker.get_events() scores = tracker.get_scores() for event, score in zip(events, scores): if "high_drift" in score.flags: print(f"Agent {event.receiver_id} showed unusual behavior") print(f" After message from: {event.sender_id}") print(f" Drift: {score.drift_delta:.3f}") ``` ### 4. Safety Research Study emergent behaviors in AI communication: ```python # Run simulation for round in range(100): sender, receiver = select_agents() result = tracker.track_event( sender_id=sender.id, receiver_id=receiver.id, sender_content=sender.generate(), receiver_content=receiver.respond(), ) # Analyze patterns summary = tracker.summary() plot_influence_network(tracker.graph, output_path="emergence.png") ``` ## Project Structure ``` TraceIQ/ ├── src/traceiq/ │ ├── __init__.py # Public API exports │ ├── models.py # Pydantic data models │ ├── tracker.py # Main InfluenceTracker class │ ├── embeddings.py # Embedding backends │ ├── scoring.py # Drift & influence calculations │ ├── metrics.py # IEEE metric computations (v0.3.0) │ ├── capabilities.py # Agent capability registry (v0.3.0) │ ├── graph.py # NetworkX graph analytics │ ├── plotting.py # Matplotlib visualizations │ ├── cli.py # Click-based CLI │ ├── export.py # CSV/JSONL export │ └── storage/ │ ├── base.py # Abstract storage interface │ ├── memory.py # In-memory backend │ └── sqlite.py # SQLite backend ├── research/ # Research experiment scripts (v0.3.0) │ ├── synthetic_simulation.py │ ├── ablation_study.py │ └── sensitivity_analysis.py ├── tests/ # Pytest test suite (116 tests) ├── examples/ │ ├── simulate_infection.py # Idea propagation simulation │ └── test_real_agents.py # Real embedding test ├── docs/ # MkDocs documentation ├── MATH.md # Mathematical framework (v0.3.0) ├── pyproject.toml # Package configuration └── README.md ``` ## API Reference ### Core Classes | Class | Description | |-------|-------------| | `InfluenceTracker` | Main class for tracking interactions | | `TrackerConfig` | Configuration options | | `InteractionEvent` | Pydantic model for events | | `ScoreResult` | Pydantic model for scores (includes IEEE metrics) | | `SummaryReport` | Aggregated metrics report | | `CapabilityRegistry` | Agent capability management (v0.3.0) | | `PropagationRiskResult` | Propagation risk over time (v0.3.0) | ### Storage Backends | Class | Description | |-------|-------------| | `MemoryStorage` | In-memory storage (default) | | `SQLiteStorage` | Persistent SQLite storage | ### Plotting Functions | Function | Description | |----------|-------------| | `plot_drift_over_time()` | Line plot of drift per agent | | `plot_influence_heatmap()` | Matrix heatmap of influence scores | | `plot_top_influencers()` | Horizontal bar chart | | `plot_influence_network()` | NetworkX graph visualization | | `plot_iqx_heatmap()` | IQx matrix heatmap (v0.3.0) | | `plot_propagation_risk_over_time()` | Spectral radius over time (v0.3.0) | | `plot_z_score_distribution()` | Z-score histogram with threshold (v0.3.0) | | `plot_top_risky_agents()` | RWI comparison chart (v0.3.0) | ## Examples Run the included examples: ```bash # Quick smoke test (uses MockEmbedder - no heavy deps) python examples/smoke.py # Simulate idea spreading through agents python examples/simulate_infection.py # Test with real sentence-transformer embeddings python examples/test_real_agents.py ``` ## Development ```bash # Install dev dependencies pip install -e ".[all,dev]" # Run linter ruff check src/ tests/ # Run tests pytest -v # Build documentation pip install mkdocs mkdocs-material mkdocs serve # Local preview at http://127.0.0.1:8000 ``` ## Dependencies **Core:** - pydantic >= 2.0 - numpy >= 1.24 - networkx >= 3.0 - click >= 8.0 - rich >= 13.0 **Optional:** - sentence-transformers >= 2.2 (`[embedding]`) - matplotlib >= 3.7 (`[plot]`) - pandas >= 2.0, scipy >= 1.10 (`[research]` - for v0.3.0 research scripts) ## License MIT License - see [LICENSE](LICENSE) for details. ## Contributing Contributions are welcome! Please feel free to submit a Pull Request. 1. Fork the repository 2. Create your feature branch (`git checkout -b feature/amazing-feature`) 3. Run tests (`pytest`) 4. Run linter (`ruff check src/ tests/`) 5. Commit your changes (`git commit -m 'Add amazing feature'`) 6. Push to the branch (`git push origin feature/amazing-feature`) 7. Open a Pull Request ## Citation If you use TraceIQ in your research, please cite: ```bibtex @software{traceiq, title = {TraceIQ: Measure AI-to-AI Influence in Multi-Agent Systems}, year = {2024}, url = {https://github.com/Anarv2104/TraceIQ} } ``` --- *Built for AI safety researchers and multi-agent system developers*
text/markdown
TraceIQ Contributors
null
null
null
MIT
agents, ai, embeddings, ieee, influence, research, security, tracking
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Science/Research", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Py...
[]
null
null
>=3.10
[]
[]
[]
[ "click>=8.0", "networkx>=3.0", "numpy>=1.24", "pydantic>=2.0", "rich>=13.0", "matplotlib>=3.7; extra == \"all\"", "pandas>=2.0; extra == \"all\"", "scipy>=1.10; extra == \"all\"", "sentence-transformers>=2.2; extra == \"all\"", "pytest>=7.0; extra == \"dev\"", "ruff>=0.1.0; extra == \"dev\"", ...
[]
[]
[]
[ "Homepage, https://github.com/Anarv2104/TraceIQ", "Documentation, https://github.com/Anarv2104/TraceIQ#readme", "Repository, https://github.com/Anarv2104/TraceIQ" ]
twine/6.1.0 CPython/3.13.7
2026-02-19T11:53:42.546600
traceiq-0.3.3.tar.gz
81,867
7e/d5/304f66b71f9d8ed90ab9ad0cb34a90b9c186ff103796b8479ae72039b1fa/traceiq-0.3.3.tar.gz
source
sdist
null
false
71df4b0364d549f7de0dac91f0dc8291
ad1b1b527bcd01eab16ecfc06cb08961828a07fcabee0d6d8b76d9cc08c29ad0
7ed5304f66b71f9d8ed90ab9ad0cb34a90b9c186ff103796b8479ae72039b1fa
null
[]
221
2.4
lakefs-sdk
1.78.0
lakeFS API
# lakefs-sdk lakeFS HTTP API This Python package is automatically generated by the [OpenAPI Generator](https://openapi-generator.tech) project: - API version: 1.0.0 - Package version: 1.78.0 - Build package: org.openapitools.codegen.languages.PythonClientCodegen ## Requirements. Python 3.7+ ## Installation & Usage ### pip install If the python package is hosted on a repository, you can install directly using: ```sh pip install git+https://github.com/treeverse/lakeFS.git ``` (you may need to run `pip` with root permission: `sudo pip install git+https://github.com/treeverse/lakeFS.git`) Then import the package: ```python import lakefs_sdk ``` ### Setuptools Install via [Setuptools](http://pypi.python.org/pypi/setuptools). ```sh python setup.py install --user ``` (or `sudo python setup.py install` to install the package for all users) Then import the package: ```python import lakefs_sdk ``` ### Tests Execute `pytest` to run the tests. ## Getting Started Please follow the [installation procedure](#installation--usage) and then run the following: ```python import time import lakefs_sdk from lakefs_sdk.rest import ApiException from pprint import pprint # Defining the host is optional and defaults to /api/v1 # See configuration.py for a list of all supported configuration parameters. configuration = lakefs_sdk.Configuration( host = "/api/v1" ) # The client must configure the authentication and authorization parameters # in accordance with the API server security policy. # Examples for each auth method are provided below, use the example that # satisfies your auth use case. # Configure HTTP basic authorization: basic_auth configuration = lakefs_sdk.Configuration( username = os.environ["USERNAME"], password = os.environ["PASSWORD"] ) # Configure API key authorization: cookie_auth configuration.api_key['cookie_auth'] = os.environ["API_KEY"] # Uncomment below to setup prefix (e.g. Bearer) for API key, if needed # configuration.api_key_prefix['cookie_auth'] = 'Bearer' # Configure API key authorization: oidc_auth configuration.api_key['oidc_auth'] = os.environ["API_KEY"] # Uncomment below to setup prefix (e.g. Bearer) for API key, if needed # configuration.api_key_prefix['oidc_auth'] = 'Bearer' # Configure API key authorization: saml_auth configuration.api_key['saml_auth'] = os.environ["API_KEY"] # Uncomment below to setup prefix (e.g. Bearer) for API key, if needed # configuration.api_key_prefix['saml_auth'] = 'Bearer' # Configure Bearer authorization (JWT): jwt_token configuration = lakefs_sdk.Configuration( access_token = os.environ["BEARER_TOKEN"] ) # Enter a context with an instance of the API client with lakefs_sdk.ApiClient(configuration) as api_client: # Create an instance of the API class api_instance = lakefs_sdk.ActionsApi(api_client) repository = 'repository_example' # str | run_id = 'run_id_example' # str | try: # get a run api_response = api_instance.get_run(repository, run_id) print("The response of ActionsApi->get_run:\n") pprint(api_response) except ApiException as e: print("Exception when calling ActionsApi->get_run: %s\n" % e) ``` ## Documentation for API Endpoints All URIs are relative to */api/v1* Class | Method | HTTP request | Description ------------ | ------------- | ------------- | ------------- *ActionsApi* | [**get_run**](https://pydocs-sdk.lakefs.io/v1.78/docs/ActionsApi.html#get_run) | **GET** /repositories/{repository}/actions/runs/{run_id} | get a run *ActionsApi* | [**get_run_hook_output**](https://pydocs-sdk.lakefs.io/v1.78/docs/ActionsApi.html#get_run_hook_output) | **GET** /repositories/{repository}/actions/runs/{run_id}/hooks/{hook_run_id}/output | get run hook output *ActionsApi* | [**list_repository_runs**](https://pydocs-sdk.lakefs.io/v1.78/docs/ActionsApi.html#list_repository_runs) | **GET** /repositories/{repository}/actions/runs | list runs *ActionsApi* | [**list_run_hooks**](https://pydocs-sdk.lakefs.io/v1.78/docs/ActionsApi.html#list_run_hooks) | **GET** /repositories/{repository}/actions/runs/{run_id}/hooks | list run hooks *AuthApi* | [**add_group_membership**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#add_group_membership) | **PUT** /auth/groups/{groupId}/members/{userId} | add group membership *AuthApi* | [**attach_policy_to_group**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#attach_policy_to_group) | **PUT** /auth/groups/{groupId}/policies/{policyId} | attach policy to group *AuthApi* | [**attach_policy_to_user**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#attach_policy_to_user) | **PUT** /auth/users/{userId}/policies/{policyId} | attach policy to user *AuthApi* | [**create_credentials**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#create_credentials) | **POST** /auth/users/{userId}/credentials | create credentials *AuthApi* | [**create_group**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#create_group) | **POST** /auth/groups | create group *AuthApi* | [**create_policy**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#create_policy) | **POST** /auth/policies | create policy *AuthApi* | [**create_user**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#create_user) | **POST** /auth/users | create user *AuthApi* | [**create_user_external_principal**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#create_user_external_principal) | **POST** /auth/users/{userId}/external/principals | attach external principal to user *AuthApi* | [**delete_credentials**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#delete_credentials) | **DELETE** /auth/users/{userId}/credentials/{accessKeyId} | delete credentials *AuthApi* | [**delete_group**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#delete_group) | **DELETE** /auth/groups/{groupId} | delete group *AuthApi* | [**delete_group_membership**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#delete_group_membership) | **DELETE** /auth/groups/{groupId}/members/{userId} | delete group membership *AuthApi* | [**delete_policy**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#delete_policy) | **DELETE** /auth/policies/{policyId} | delete policy *AuthApi* | [**delete_user**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#delete_user) | **DELETE** /auth/users/{userId} | delete user *AuthApi* | [**delete_user_external_principal**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#delete_user_external_principal) | **DELETE** /auth/users/{userId}/external/principals | delete external principal from user *AuthApi* | [**detach_policy_from_group**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#detach_policy_from_group) | **DELETE** /auth/groups/{groupId}/policies/{policyId} | detach policy from group *AuthApi* | [**detach_policy_from_user**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#detach_policy_from_user) | **DELETE** /auth/users/{userId}/policies/{policyId} | detach policy from user *AuthApi* | [**external_principal_login**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#external_principal_login) | **POST** /auth/external/principal/login | perform a login using an external authenticator *AuthApi* | [**get_credentials**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#get_credentials) | **GET** /auth/users/{userId}/credentials/{accessKeyId} | get credentials *AuthApi* | [**get_current_user**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#get_current_user) | **GET** /user | get current user *AuthApi* | [**get_external_principal**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#get_external_principal) | **GET** /auth/external/principals | describe external principal by id *AuthApi* | [**get_group**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#get_group) | **GET** /auth/groups/{groupId} | get group *AuthApi* | [**get_group_acl**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#get_group_acl) | **GET** /auth/groups/{groupId}/acl | get ACL of group *AuthApi* | [**get_policy**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#get_policy) | **GET** /auth/policies/{policyId} | get policy *AuthApi* | [**get_token_from_mailbox**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#get_token_from_mailbox) | **GET** /auth/get-token/mailboxes/{mailbox} | receive the token after user has authenticated on redirect URL. *AuthApi* | [**get_token_redirect**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#get_token_redirect) | **GET** /auth/get-token/start | start acquiring a token by logging in on a browser *AuthApi* | [**get_user**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#get_user) | **GET** /auth/users/{userId} | get user *AuthApi* | [**list_group_members**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#list_group_members) | **GET** /auth/groups/{groupId}/members | list group members *AuthApi* | [**list_group_policies**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#list_group_policies) | **GET** /auth/groups/{groupId}/policies | list group policies *AuthApi* | [**list_groups**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#list_groups) | **GET** /auth/groups | list groups *AuthApi* | [**list_policies**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#list_policies) | **GET** /auth/policies | list policies *AuthApi* | [**list_user_credentials**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#list_user_credentials) | **GET** /auth/users/{userId}/credentials | list user credentials *AuthApi* | [**list_user_external_principals**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#list_user_external_principals) | **GET** /auth/users/{userId}/external/principals/ls | list user external policies attached to a user *AuthApi* | [**list_user_groups**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#list_user_groups) | **GET** /auth/users/{userId}/groups | list user groups *AuthApi* | [**list_user_policies**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#list_user_policies) | **GET** /auth/users/{userId}/policies | list user policies *AuthApi* | [**list_users**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#list_users) | **GET** /auth/users | list users *AuthApi* | [**login**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#login) | **POST** /auth/login | perform a login *AuthApi* | [**oauth_callback**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#oauth_callback) | **GET** /oidc/callback | *AuthApi* | [**release_token_to_mailbox**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#release_token_to_mailbox) | **GET** /auth/get-token/release-token/{loginRequestToken} | release a token for the current (authenticated) user to the mailbox of this login request. *AuthApi* | [**set_group_acl**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#set_group_acl) | **POST** /auth/groups/{groupId}/acl | set ACL of group *AuthApi* | [**update_policy**](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthApi.html#update_policy) | **PUT** /auth/policies/{policyId} | update policy *BranchesApi* | [**cherry_pick**](https://pydocs-sdk.lakefs.io/v1.78/docs/BranchesApi.html#cherry_pick) | **POST** /repositories/{repository}/branches/{branch}/cherry-pick | Replay the changes from the given commit on the branch *BranchesApi* | [**create_branch**](https://pydocs-sdk.lakefs.io/v1.78/docs/BranchesApi.html#create_branch) | **POST** /repositories/{repository}/branches | create branch *BranchesApi* | [**delete_branch**](https://pydocs-sdk.lakefs.io/v1.78/docs/BranchesApi.html#delete_branch) | **DELETE** /repositories/{repository}/branches/{branch} | delete branch *BranchesApi* | [**diff_branch**](https://pydocs-sdk.lakefs.io/v1.78/docs/BranchesApi.html#diff_branch) | **GET** /repositories/{repository}/branches/{branch}/diff | diff branch *BranchesApi* | [**get_branch**](https://pydocs-sdk.lakefs.io/v1.78/docs/BranchesApi.html#get_branch) | **GET** /repositories/{repository}/branches/{branch} | get branch *BranchesApi* | [**list_branches**](https://pydocs-sdk.lakefs.io/v1.78/docs/BranchesApi.html#list_branches) | **GET** /repositories/{repository}/branches | list branches *BranchesApi* | [**reset_branch**](https://pydocs-sdk.lakefs.io/v1.78/docs/BranchesApi.html#reset_branch) | **PUT** /repositories/{repository}/branches/{branch} | reset branch *BranchesApi* | [**revert_branch**](https://pydocs-sdk.lakefs.io/v1.78/docs/BranchesApi.html#revert_branch) | **POST** /repositories/{repository}/branches/{branch}/revert | revert *CommitsApi* | [**commit**](https://pydocs-sdk.lakefs.io/v1.78/docs/CommitsApi.html#commit) | **POST** /repositories/{repository}/branches/{branch}/commits | create commit *CommitsApi* | [**get_commit**](https://pydocs-sdk.lakefs.io/v1.78/docs/CommitsApi.html#get_commit) | **GET** /repositories/{repository}/commits/{commitId} | get commit *ConfigApi* | [**get_config**](https://pydocs-sdk.lakefs.io/v1.78/docs/ConfigApi.html#get_config) | **GET** /config | *ExperimentalApi* | [**abort_presign_multipart_upload**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#abort_presign_multipart_upload) | **DELETE** /repositories/{repository}/branches/{branch}/staging/pmpu/{uploadId} | Abort a presign multipart upload *ExperimentalApi* | [**commit_async**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#commit_async) | **POST** /repositories/{repository}/branches/{branch}/commits/async | create commit asynchronously *ExperimentalApi* | [**commit_async_status**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#commit_async_status) | **GET** /repositories/{repository}/branches/{branch}/commits/async/{id}/status | get status of async commit operation *ExperimentalApi* | [**complete_presign_multipart_upload**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#complete_presign_multipart_upload) | **PUT** /repositories/{repository}/branches/{branch}/staging/pmpu/{uploadId} | Complete a presign multipart upload request *ExperimentalApi* | [**create_presign_multipart_upload**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#create_presign_multipart_upload) | **POST** /repositories/{repository}/branches/{branch}/staging/pmpu | Initiate a multipart upload *ExperimentalApi* | [**create_pull_request**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#create_pull_request) | **POST** /repositories/{repository}/pulls | create pull request *ExperimentalApi* | [**create_user_external_principal**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#create_user_external_principal) | **POST** /auth/users/{userId}/external/principals | attach external principal to user *ExperimentalApi* | [**delete_user_external_principal**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#delete_user_external_principal) | **DELETE** /auth/users/{userId}/external/principals | delete external principal from user *ExperimentalApi* | [**external_principal_login**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#external_principal_login) | **POST** /auth/external/principal/login | perform a login using an external authenticator *ExperimentalApi* | [**get_external_principal**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#get_external_principal) | **GET** /auth/external/principals | describe external principal by id *ExperimentalApi* | [**get_license**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#get_license) | **GET** /license | *ExperimentalApi* | [**get_pull_request**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#get_pull_request) | **GET** /repositories/{repository}/pulls/{pull_request} | get pull request *ExperimentalApi* | [**get_token_from_mailbox**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#get_token_from_mailbox) | **GET** /auth/get-token/mailboxes/{mailbox} | receive the token after user has authenticated on redirect URL. *ExperimentalApi* | [**get_token_redirect**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#get_token_redirect) | **GET** /auth/get-token/start | start acquiring a token by logging in on a browser *ExperimentalApi* | [**hard_reset_branch**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#hard_reset_branch) | **PUT** /repositories/{repository}/branches/{branch}/hard_reset | hard reset branch *ExperimentalApi* | [**list_pull_requests**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#list_pull_requests) | **GET** /repositories/{repository}/pulls | list pull requests *ExperimentalApi* | [**list_user_external_principals**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#list_user_external_principals) | **GET** /auth/users/{userId}/external/principals/ls | list user external policies attached to a user *ExperimentalApi* | [**merge_into_branch_async**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#merge_into_branch_async) | **POST** /repositories/{repository}/refs/{sourceRef}/merge/{destinationBranch}/async | merge references asynchronously *ExperimentalApi* | [**merge_into_branch_async_status**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#merge_into_branch_async_status) | **GET** /repositories/{repository}/refs/{sourceRef}/merge/{destinationBranch}/async/{id}/status | get status of async merge operation *ExperimentalApi* | [**merge_pull_request**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#merge_pull_request) | **PUT** /repositories/{repository}/pulls/{pull_request}/merge | merge pull request *ExperimentalApi* | [**release_token_to_mailbox**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#release_token_to_mailbox) | **GET** /auth/get-token/release-token/{loginRequestToken} | release a token for the current (authenticated) user to the mailbox of this login request. *ExperimentalApi* | [**sts_login**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#sts_login) | **POST** /sts/login | perform a login with STS *ExperimentalApi* | [**update_object_user_metadata**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#update_object_user_metadata) | **PUT** /repositories/{repository}/branches/{branch}/objects/stat/user_metadata | rewrite (all) object metadata *ExperimentalApi* | [**update_pull_request**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#update_pull_request) | **PATCH** /repositories/{repository}/pulls/{pull_request} | update pull request *ExperimentalApi* | [**upload_part**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#upload_part) | **PUT** /repositories/{repository}/branches/{branch}/staging/pmpu/{uploadId}/parts/{partNumber} | *ExperimentalApi* | [**upload_part_copy**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExperimentalApi.html#upload_part_copy) | **PUT** /repositories/{repository}/branches/{branch}/staging/pmpu/{uploadId}/parts/{partNumber}/copy | *ExternalApi* | [**create_user_external_principal**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExternalApi.html#create_user_external_principal) | **POST** /auth/users/{userId}/external/principals | attach external principal to user *ExternalApi* | [**delete_user_external_principal**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExternalApi.html#delete_user_external_principal) | **DELETE** /auth/users/{userId}/external/principals | delete external principal from user *ExternalApi* | [**external_principal_login**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExternalApi.html#external_principal_login) | **POST** /auth/external/principal/login | perform a login using an external authenticator *ExternalApi* | [**get_external_principal**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExternalApi.html#get_external_principal) | **GET** /auth/external/principals | describe external principal by id *ExternalApi* | [**list_user_external_principals**](https://pydocs-sdk.lakefs.io/v1.78/docs/ExternalApi.html#list_user_external_principals) | **GET** /auth/users/{userId}/external/principals/ls | list user external policies attached to a user *HealthCheckApi* | [**health_check**](https://pydocs-sdk.lakefs.io/v1.78/docs/HealthCheckApi.html#health_check) | **GET** /healthcheck | *ImportApi* | [**import_cancel**](https://pydocs-sdk.lakefs.io/v1.78/docs/ImportApi.html#import_cancel) | **DELETE** /repositories/{repository}/branches/{branch}/import | cancel ongoing import *ImportApi* | [**import_start**](https://pydocs-sdk.lakefs.io/v1.78/docs/ImportApi.html#import_start) | **POST** /repositories/{repository}/branches/{branch}/import | import data from object store *ImportApi* | [**import_status**](https://pydocs-sdk.lakefs.io/v1.78/docs/ImportApi.html#import_status) | **GET** /repositories/{repository}/branches/{branch}/import | get import status *InternalApi* | [**create_branch_protection_rule_preflight**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#create_branch_protection_rule_preflight) | **GET** /repositories/{repository}/branch_protection/set_allowed | *InternalApi* | [**create_commit_record**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#create_commit_record) | **POST** /repositories/{repository}/commits | create commit record *InternalApi* | [**create_symlink_file**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#create_symlink_file) | **POST** /repositories/{repository}/refs/{branch}/symlink | creates symlink files corresponding to the given directory *InternalApi* | [**delete_repository_metadata**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#delete_repository_metadata) | **DELETE** /repositories/{repository}/metadata | delete repository metadata *InternalApi* | [**dump_refs**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#dump_refs) | **PUT** /repositories/{repository}/refs/dump | Dump repository refs (tags, commits, branches) to object store Deprecated: a new API will introduce long running operations *InternalApi* | [**get_auth_capabilities**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#get_auth_capabilities) | **GET** /auth/capabilities | list authentication capabilities supported *InternalApi* | [**get_garbage_collection_config**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#get_garbage_collection_config) | **GET** /config/garbage-collection | *InternalApi* | [**get_lake_fs_version**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#get_lake_fs_version) | **GET** /config/version | *InternalApi* | [**get_metadata_object**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#get_metadata_object) | **GET** /repositories/{repository}/metadata/object/{type}/{object_id} | return a lakeFS metadata object by ID *InternalApi* | [**get_setup_state**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#get_setup_state) | **GET** /setup_lakefs | check if the lakeFS installation is already set up *InternalApi* | [**get_storage_config**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#get_storage_config) | **GET** /config/storage | *InternalApi* | [**get_usage_report_summary**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#get_usage_report_summary) | **GET** /usage-report/summary | get usage report summary *InternalApi* | [**internal_create_branch_protection_rule**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#internal_create_branch_protection_rule) | **POST** /repositories/{repository}/branch_protection | *InternalApi* | [**internal_delete_branch_protection_rule**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#internal_delete_branch_protection_rule) | **DELETE** /repositories/{repository}/branch_protection | *InternalApi* | [**internal_delete_garbage_collection_rules**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#internal_delete_garbage_collection_rules) | **DELETE** /repositories/{repository}/gc/rules | *InternalApi* | [**internal_get_branch_protection_rules**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#internal_get_branch_protection_rules) | **GET** /repositories/{repository}/branch_protection | get branch protection rules *InternalApi* | [**internal_get_garbage_collection_rules**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#internal_get_garbage_collection_rules) | **GET** /repositories/{repository}/gc/rules | *InternalApi* | [**internal_set_garbage_collection_rules**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#internal_set_garbage_collection_rules) | **POST** /repositories/{repository}/gc/rules | *InternalApi* | [**post_stats_events**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#post_stats_events) | **POST** /statistics | post stats events, this endpoint is meant for internal use only *InternalApi* | [**prepare_garbage_collection_commits**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#prepare_garbage_collection_commits) | **POST** /repositories/{repository}/gc/prepare_commits | save lists of active commits for garbage collection *InternalApi* | [**prepare_garbage_collection_commits_async**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#prepare_garbage_collection_commits_async) | **POST** /repositories/{repository}/gc/prepare_commits/async | prepare gc commits *InternalApi* | [**prepare_garbage_collection_commits_status**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#prepare_garbage_collection_commits_status) | **GET** /repositories/{repository}/gc/prepare_commits/status | get status of prepare gc commits operation *InternalApi* | [**prepare_garbage_collection_uncommitted**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#prepare_garbage_collection_uncommitted) | **POST** /repositories/{repository}/gc/prepare_uncommited | save repository uncommitted metadata for garbage collection *InternalApi* | [**restore_refs**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#restore_refs) | **PUT** /repositories/{repository}/refs/restore | Restore repository refs (tags, commits, branches) from object store. Deprecated: a new API will introduce long running operations *InternalApi* | [**set_garbage_collection_rules_preflight**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#set_garbage_collection_rules_preflight) | **GET** /repositories/{repository}/gc/rules/set_allowed | *InternalApi* | [**set_repository_metadata**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#set_repository_metadata) | **POST** /repositories/{repository}/metadata | set repository metadata *InternalApi* | [**setup**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#setup) | **POST** /setup_lakefs | setup lakeFS and create a first user *InternalApi* | [**setup_comm_prefs**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#setup_comm_prefs) | **POST** /setup_comm_prefs | setup communications preferences *InternalApi* | [**stage_object**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#stage_object) | **PUT** /repositories/{repository}/branches/{branch}/objects | stage an object&#39;s metadata for the given branch *InternalApi* | [**upload_object_preflight**](https://pydocs-sdk.lakefs.io/v1.78/docs/InternalApi.html#upload_object_preflight) | **GET** /repositories/{repository}/branches/{branch}/objects/stage_allowed | *LicenseApi* | [**get_license**](https://pydocs-sdk.lakefs.io/v1.78/docs/LicenseApi.html#get_license) | **GET** /license | *MetadataApi* | [**get_meta_range**](https://pydocs-sdk.lakefs.io/v1.78/docs/MetadataApi.html#get_meta_range) | **GET** /repositories/{repository}/metadata/meta_range/{meta_range} | return URI to a meta-range file *MetadataApi* | [**get_range**](https://pydocs-sdk.lakefs.io/v1.78/docs/MetadataApi.html#get_range) | **GET** /repositories/{repository}/metadata/range/{range} | return URI to a range file *ObjectsApi* | [**copy_object**](https://pydocs-sdk.lakefs.io/v1.78/docs/ObjectsApi.html#copy_object) | **POST** /repositories/{repository}/branches/{branch}/objects/copy | create a copy of an object *ObjectsApi* | [**delete_object**](https://pydocs-sdk.lakefs.io/v1.78/docs/ObjectsApi.html#delete_object) | **DELETE** /repositories/{repository}/branches/{branch}/objects | delete object. Missing objects will not return a NotFound error. *ObjectsApi* | [**delete_objects**](https://pydocs-sdk.lakefs.io/v1.78/docs/ObjectsApi.html#delete_objects) | **POST** /repositories/{repository}/branches/{branch}/objects/delete | delete objects. Missing objects will not return a NotFound error. *ObjectsApi* | [**get_object**](https://pydocs-sdk.lakefs.io/v1.78/docs/ObjectsApi.html#get_object) | **GET** /repositories/{repository}/refs/{ref}/objects | get object content *ObjectsApi* | [**get_underlying_properties**](https://pydocs-sdk.lakefs.io/v1.78/docs/ObjectsApi.html#get_underlying_properties) | **GET** /repositories/{repository}/refs/{ref}/objects/underlyingProperties | get object properties on underlying storage *ObjectsApi* | [**head_object**](https://pydocs-sdk.lakefs.io/v1.78/docs/ObjectsApi.html#head_object) | **HEAD** /repositories/{repository}/refs/{ref}/objects | check if object exists *ObjectsApi* | [**list_objects**](https://pydocs-sdk.lakefs.io/v1.78/docs/ObjectsApi.html#list_objects) | **GET** /repositories/{repository}/refs/{ref}/objects/ls | list objects under a given prefix *ObjectsApi* | [**stat_object**](https://pydocs-sdk.lakefs.io/v1.78/docs/ObjectsApi.html#stat_object) | **GET** /repositories/{repository}/refs/{ref}/objects/stat | get object metadata *ObjectsApi* | [**update_object_user_metadata**](https://pydocs-sdk.lakefs.io/v1.78/docs/ObjectsApi.html#update_object_user_metadata) | **PUT** /repositories/{repository}/branches/{branch}/objects/stat/user_metadata | rewrite (all) object metadata *ObjectsApi* | [**upload_object**](https://pydocs-sdk.lakefs.io/v1.78/docs/ObjectsApi.html#upload_object) | **POST** /repositories/{repository}/branches/{branch}/objects | *PullsApi* | [**create_pull_request**](https://pydocs-sdk.lakefs.io/v1.78/docs/PullsApi.html#create_pull_request) | **POST** /repositories/{repository}/pulls | create pull request *PullsApi* | [**get_pull_request**](https://pydocs-sdk.lakefs.io/v1.78/docs/PullsApi.html#get_pull_request) | **GET** /repositories/{repository}/pulls/{pull_request} | get pull request *PullsApi* | [**list_pull_requests**](https://pydocs-sdk.lakefs.io/v1.78/docs/PullsApi.html#list_pull_requests) | **GET** /repositories/{repository}/pulls | list pull requests *PullsApi* | [**merge_pull_request**](https://pydocs-sdk.lakefs.io/v1.78/docs/PullsApi.html#merge_pull_request) | **PUT** /repositories/{repository}/pulls/{pull_request}/merge | merge pull request *PullsApi* | [**update_pull_request**](https://pydocs-sdk.lakefs.io/v1.78/docs/PullsApi.html#update_pull_request) | **PATCH** /repositories/{repository}/pulls/{pull_request} | update pull request *RefsApi* | [**diff_refs**](https://pydocs-sdk.lakefs.io/v1.78/docs/RefsApi.html#diff_refs) | **GET** /repositories/{repository}/refs/{leftRef}/diff/{rightRef} | diff references *RefsApi* | [**find_merge_base**](https://pydocs-sdk.lakefs.io/v1.78/docs/RefsApi.html#find_merge_base) | **GET** /repositories/{repository}/refs/{sourceRef}/merge/{destinationBranch} | find the merge base for 2 references *RefsApi* | [**log_commits**](https://pydocs-sdk.lakefs.io/v1.78/docs/RefsApi.html#log_commits) | **GET** /repositories/{repository}/refs/{ref}/commits | get commit log from ref. If both objects and prefixes are empty, return all commits. *RefsApi* | [**merge_into_branch**](https://pydocs-sdk.lakefs.io/v1.78/docs/RefsApi.html#merge_into_branch) | **POST** /repositories/{repository}/refs/{sourceRef}/merge/{destinationBranch} | merge references *RemotesApi* | [**pull_iceberg_table**](https://pydocs-sdk.lakefs.io/v1.78/docs/RemotesApi.html#pull_iceberg_table) | **POST** /iceberg/remotes/{catalog}/pull | take a table previously pushed from lakeFS into a remote catalog, and pull its state back into the originating lakeFS repository *RemotesApi* | [**push_iceberg_table**](https://pydocs-sdk.lakefs.io/v1.78/docs/RemotesApi.html#push_iceberg_table) | **POST** /iceberg/remotes/{catalog}/push | register existing lakeFS table in remote catalog *RepositoriesApi* | [**create_repository**](https://pydocs-sdk.lakefs.io/v1.78/docs/RepositoriesApi.html#create_repository) | **POST** /repositories | create repository *RepositoriesApi* | [**delete_gc_rules**](https://pydocs-sdk.lakefs.io/v1.78/docs/RepositoriesApi.html#delete_gc_rules) | **DELETE** /repositories/{repository}/settings/gc_rules | *RepositoriesApi* | [**delete_repository**](https://pydocs-sdk.lakefs.io/v1.78/docs/RepositoriesApi.html#delete_repository) | **DELETE** /repositories/{repository} | delete repository *RepositoriesApi* | [**dump_status**](https://pydocs-sdk.lakefs.io/v1.78/docs/RepositoriesApi.html#dump_status) | **GET** /repositories/{repository}/dump | Status of a repository dump task *RepositoriesApi* | [**dump_submit**](https://pydocs-sdk.lakefs.io/v1.78/docs/RepositoriesApi.html#dump_submit) | **POST** /repositories/{repository}/dump | Backup the repository metadata (tags, commits, branches) and save the backup to the object store. *RepositoriesApi* | [**get_branch_protection_rules**](https://pydocs-sdk.lakefs.io/v1.78/docs/RepositoriesApi.html#get_branch_protection_rules) | **GET** /repositories/{repository}/settings/branch_protection | get branch protection rules *RepositoriesApi* | [**get_gc_rules**](https://pydocs-sdk.lakefs.io/v1.78/docs/RepositoriesApi.html#get_gc_rules) | **GET** /repositories/{repository}/settings/gc_rules | get repository GC rules *RepositoriesApi* | [**get_repository**](https://pydocs-sdk.lakefs.io/v1.78/docs/RepositoriesApi.html#get_repository) | **GET** /repositories/{repository} | get repository *RepositoriesApi* | [**get_repository_metadata**](https://pydocs-sdk.lakefs.io/v1.78/docs/RepositoriesApi.html#get_repository_metadata) | **GET** /repositories/{repository}/metadata | get repository metadata *RepositoriesApi* | [**list_repositories**](https://pydocs-sdk.lakefs.io/v1.78/docs/RepositoriesApi.html#list_repositories) | **GET** /repositories | list repositories *RepositoriesApi* | [**restore_status**](https://pydocs-sdk.lakefs.io/v1.78/docs/RepositoriesApi.html#restore_status) | **GET** /repositories/{repository}/restore | Status of a restore request *RepositoriesApi* | [**restore_submit**](https://pydocs-sdk.lakefs.io/v1.78/docs/RepositoriesApi.html#restore_submit) | **POST** /repositories/{repository}/restore | Restore repository from a dump in the object store *RepositoriesApi* | [**set_branch_protection_rules**](https://pydocs-sdk.lakefs.io/v1.78/docs/RepositoriesApi.html#set_branch_protection_rules) | **PUT** /repositories/{repository}/settings/branch_protection | *RepositoriesApi* | [**set_gc_rules**](https://pydocs-sdk.lakefs.io/v1.78/docs/RepositoriesApi.html#set_gc_rules) | **PUT** /repositories/{repository}/settings/gc_rules | *StagingApi* | [**get_physical_address**](https://pydocs-sdk.lakefs.io/v1.78/docs/StagingApi.html#get_physical_address) | **GET** /repositories/{repository}/branches/{branch}/staging/backing | generate an address to which the client can upload an object *StagingApi* | [**link_physical_address**](https://pydocs-sdk.lakefs.io/v1.78/docs/StagingApi.html#link_physical_address) | **PUT** /repositories/{repository}/branches/{branch}/staging/backing | associate staging on this physical address with a path *TagsApi* | [**create_tag**](https://pydocs-sdk.lakefs.io/v1.78/docs/TagsApi.html#create_tag) | **POST** /repositories/{repository}/tags | create tag *TagsApi* | [**delete_tag**](https://pydocs-sdk.lakefs.io/v1.78/docs/TagsApi.html#delete_tag) | **DELETE** /repositories/{repository}/tags/{tag} | delete tag *TagsApi* | [**get_tag**](https://pydocs-sdk.lakefs.io/v1.78/docs/TagsApi.html#get_tag) | **GET** /repositories/{repository}/tags/{tag} | get tag *TagsApi* | [**list_tags**](https://pydocs-sdk.lakefs.io/v1.78/docs/TagsApi.html#list_tags) | **GET** /repositories/{repository}/tags | list tags ## Documentation For Models - [ACL](https://pydocs-sdk.lakefs.io/v1.78/docs/ACL.html) - [AbortPresignMultipartUpload](https://pydocs-sdk.lakefs.io/v1.78/docs/AbortPresignMultipartUpload.html) - [AccessKeyCredentials](https://pydocs-sdk.lakefs.io/v1.78/docs/AccessKeyCredentials.html) - [ActionRun](https://pydocs-sdk.lakefs.io/v1.78/docs/ActionRun.html) - [ActionRunList](https://pydocs-sdk.lakefs.io/v1.78/docs/ActionRunList.html) - [AsyncTaskStatus](https://pydocs-sdk.lakefs.io/v1.78/docs/AsyncTaskStatus.html) - [AuthCapabilities](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthCapabilities.html) - [AuthenticationToken](https://pydocs-sdk.lakefs.io/v1.78/docs/AuthenticationToken.html) - [BranchCreation](https://pydocs-sdk.lakefs.io/v1.78/docs/BranchCreation.html) - [BranchProtectionRule](https://pydocs-sdk.lakefs.io/v1.78/docs/BranchProtectionRule.html) - [CapabilitiesConfig](https://pydocs-sdk.lakefs.io/v1.78/docs/CapabilitiesConfig.html) - [CherryPickCreation](https://pydocs-sdk.lakefs.io/v1.78/docs/CherryPickCreation.html) - [CommPrefsInput](https://pydocs-sdk.lakefs.io/v1.78/docs/CommPrefsInput.html) - [Commit](https://pydocs-sdk.lakefs.io/v1.78/docs/Commit.html) - [CommitAsyncStatus](https://pydocs-sdk.lakefs.io/v1.78/docs/CommitAsyncStatus.html) - [CommitCreation](https://pydocs-sdk.lakefs.io/v1.78/docs/CommitCreation.html) - [CommitList](https://pydocs-sdk.lakefs.io/v1.78/docs/CommitList.html) - [CommitOverrides](https://pydocs-sdk.lakefs.io/v1.78/docs/CommitOverrides.html) - [CommitRecordCreation](https://pydocs-sdk.lakefs.io/v1.78/docs/CommitRecordCreation.html) - [CompletePresignMultipartUpload](https://pydocs-sdk.lakefs.io/v1.78/docs/CompletePresignMultipartUpload.html) - [Config](https://pydocs-sdk.lakefs.io/v1.78/docs/Config.html) - [CopyPartSource](https://pydocs-sdk.lakefs.io/v1.78/docs/CopyPartSource.html) - [Credentials](https://pydocs-sdk.lakefs.io/v1.78/docs/Credentials.html) - [CredentialsList](https://pydocs-sdk.lakefs.io/v1.78/docs/CredentialsList.html) - [CredentialsWithSecret](https://pydocs-sdk.lakefs.io/v1.78/docs/CredentialsWithSecret.html) - [CurrentUser](https://pydocs-sdk.lakefs.io/v1.78/docs/CurrentUser.html) - [CustomViewer](https://pydocs-sdk.lakefs.io/v1.78/docs/CustomViewer.html) - [Diff](https://pydocs-sdk.lakefs.io/v1.78/docs/Diff.html) - [DiffList](https://pydocs-sdk.lakefs.io/v1.78/docs/DiffList.html) - [DiffObjectStat](https://pydocs-sdk.lakefs.io/v1.78/docs/DiffObjectStat.html) - [Error](https://pydocs-sdk.lakefs.io/v1.78/docs/Error.html) - [ErrorNoACL](https://pydocs-sdk.lakefs.io/v1.78/docs/ErrorNoACL.html) - [ExternalLoginInformation](https://pydocs-sdk.lakefs.io/v1.78/docs/ExternalLoginInformation.html) - [ExternalPrincipal](https://pydocs-sdk.lakefs.io/v1.78/docs/ExternalPrincipal.html) - [ExternalPrincipalCreation](https://pydocs-sdk.lakefs.io/v1.78/docs/ExternalPrincipalCreation.html) - [ExternalPrincipalList](https://pydocs-sdk.lakefs.io/v1.78/docs/ExternalPrincipalList.html) - [FindMergeBaseResult](https://pydocs-sdk.lakefs.io/v1.78/docs/FindMergeBaseResult.html) - [GarbageCollectionConfig](https://pydocs-sdk.lakefs.io/v1.78/docs/GarbageCollectionConfig.html) - [GarbageCollectionPrepareResponse](https://pydocs-sdk.lakefs.io/v1.78/docs/GarbageCollectionPrepareResponse.html) - [GarbageCollectionRule](https://pydocs-sdk.lakefs.io/v1.78/docs/GarbageCollectionRule.html) - [GarbageCollectionRules](https://pydocs-sdk.lakefs.io/v1.78/docs/GarbageCollectionRules.html) - [Group](https://pydocs-sdk.lakefs.io/v1.78/docs/Group.html) - [GroupCreation](https://pydocs-sdk.lakefs.io/v1.78/docs/GroupCreation.html) - [GroupList](https://pydocs-sdk.lakefs.io/v1.78/docs/GroupList.html) - [HookRun](https://pydocs-sdk.lakefs.io/v1.78/docs/HookRun.html) - [HookRunList](https://pydocs-sdk.lakefs.io/v1.78/docs/HookRunList.html) - [IcebergLocalTable](https://pydocs-sdk.lakefs.io/v1.78/docs/IcebergLocalTable.html) - [IcebergPullRequest](https://pydocs-sdk.lakefs.io/v1.78/docs/Iceb
text/markdown
Treeverse
services@treeverse.io
null
null
Apache 2.0
OpenAPI, OpenAPI-Generator, lakeFS API
[]
[]
https://github.com/treeverse/lakeFS/tree/master/clients/python
null
>=3.6
[]
[]
[]
[ "urllib3<3.0.0,>=1.25.3", "python-dateutil", "pydantic>=1.10.5", "aenum" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-19T11:53:31.786458
lakefs_sdk-1.78.0.tar.gz
138,267
87/35/66efb07b66b87860225808b15483eef992619eae03d692ec357612fe54b8/lakefs_sdk-1.78.0.tar.gz
source
sdist
null
false
99032d185522bda20e3696527ad50e49
1ef85927629ca6ecb5cb01910cf16dd0ca5a91375866ddcb1ecd2293ceac9d25
873566efb07b66b87860225808b15483eef992619eae03d692ec357612fe54b8
null
[]
379,584
2.4
agentstateprotocol
0.1.0
Checkpointing and recovery protocol for AI agents
# AgentStateProtocol Checkpointing and recovery protocol for AI agents. `AgentStateProtocol` lets an agent save state at each reasoning step, branch into alternatives, and roll back safely after failures. ## Install ```bash pip install agentstateprotocol ``` ## Quick Start ```python from agentstateprotocol import AgentStateProtocol agent = AgentStateProtocol("my-agent") agent.checkpoint( state={"task": "summarize quarterly report", "stage": "parsed"}, metadata={"confidence": 0.91}, description="Initial parse", logic_step="parse_input", ) agent.checkpoint( state={"task": "summarize quarterly report", "stage": "drafted"}, metadata={"confidence": 0.86}, description="Draft generated", logic_step="draft_summary", ) # Roll back one step if needed agent.rollback() ``` ## Core Operations - `agent.checkpoint(...)`: save a state snapshot - `agent.rollback(...)`: restore a previous checkpoint - `agent.branch(name)`: start an alternate reasoning path - `agent.switch_branch(name)`: move between branches - `agent.merge(source_branch)`: merge branch outcomes - `agent.history()`: inspect checkpoint timeline - `agent.visualize_tree()`: show decision tree ## Decorators ```python from agentstateprotocol.decorators import agentstateprotocol_step @agentstateprotocol_step("analyze") def analyze(state): return {"result": "ok", **state} ``` ## Storage ```python from agentstateprotocol.storage import FileSystemStorage, SQLiteStorage file_storage = FileSystemStorage(".agentstateprotocol") sqlite_storage = SQLiteStorage(".agentstateprotocol/agentstateprotocol.db") ``` ## CLI ```bash agentstateprotocol demo agentstateprotocol log agentstateprotocol tree agentstateprotocol branches agentstateprotocol diff <checkpoint_a> <checkpoint_b> agentstateprotocol metrics ``` ## Project - Repository: https://github.com/ekessh/agentstateprotocol - License: MIT
text/markdown
AgentStateProtocol Contributors
null
null
null
MIT
ai, agents, checkpointing, recovery, state-management, llm, reasoning, rollback, branching
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: P...
[]
null
null
>=3.9
[]
[]
[]
[ "pytest>=7.0; extra == \"dev\"", "pytest-cov>=4.0; extra == \"dev\"" ]
[]
[]
[]
[ "Homepage, https://github.com/ekessh/agentstateprotocol", "Documentation, https://agentstateprotocol.readthedocs.io", "Repository, https://github.com/ekessh/agentstateprotocol", "Issues, https://github.com/ekessh/agentstateprotocol/issues" ]
twine/6.2.0 CPython/3.13.5
2026-02-19T11:51:22.015718
agentstateprotocol-0.1.0.tar.gz
22,587
5b/77/aea42505dc09434b7537f91d94339b45e04ce0e0bce4362251069f5c78bc/agentstateprotocol-0.1.0.tar.gz
source
sdist
null
false
00219a68a0d6d842bba3f6b2b3427a29
0b93f2d6ac8f88da26ffbd6d8be6e837ad8cb6c4b4c6966c3f3be40ec25c9721
5b77aea42505dc09434b7537f91d94339b45e04ce0e0bce4362251069f5c78bc
null
[ "LICENSE" ]
258
2.4
librelane
3.0.0.dev52
An infrastructure for implementing chip design flows
<h1 align="center">LibreLane</h1> <p align="center"> <a href="https://opensource.org/licenses/Apache-2.0"><img src="https://img.shields.io/badge/License-Apache%202.0-blue.svg" alt="License: Apache 2.0"/></a> <a href="https://www.python.org"><img src="https://img.shields.io/badge/Python-3.10-3776AB.svg?style=flat&logo=python&logoColor=white" alt="Python ≥3.10" /></a> <a href="https://github.com/psf/black"><img src="https://img.shields.io/badge/code%20style-black-000000.svg" alt="Code Style: black"/></a> <a href="https://mypy-lang.org/"><img src="https://www.mypy-lang.org/static/mypy_badge.svg" alt="Checked with mypy"/></a> <a href="https://nixos.org/"><img src="https://img.shields.io/static/v1?logo=nixos&logoColor=white&label=&message=Built%20with%20Nix&color=41439a" alt="Built with Nix"/></a> </p> <p align="center"> <a href="https://colab.research.google.com/github/librelane/librelane/blob/main/notebook.ipynb"><img src="https://colab.research.google.com/assets/colab-badge.svg" alt="Open in Colab"></a> <a href="https://librelane.readthedocs.io/"><img src="https://readthedocs.org/projects/librelane/badge/?version=latest" alt="Documentation Build Status Badge"/></a> <a href="https://fossi-chat.org"><img src="https://img.shields.io/badge/Community-FOSSi%20Chat-1bb378?logo=element" alt="Invite to FOSSi Chat"/></a> </p> LibreLane is an ASIC infrastructure library based on several components including OpenROAD, Yosys, Magic, Netgen, CVC, KLayout and a number of custom scripts for design exploration and optimization, currently developed and maintained under the stewardship of the [FOSSi Foundation](https://fossi-foundation.org). A reference flow, "Classic", performs all ASIC implementation steps from RTL all the way down to GDSII. You can find the documentation [here](https://librelane.readthedocs.io/en/latest/getting_started/) to get started. You can discuss LibreLane in the [FOSSi Chat Matrix Server](https://fossi-chat.org). ## Try it out You can try LibreLane right in your browser, free-of-charge, using Google Colaboratory by following [**this link**](https://colab.research.google.com/github/librelane/librelane/blob/main/notebook.ipynb). ## Installation You'll need the following: * Python **3.10** or higher with PIP, Venv and Tkinter ### Nix (Recommended) Works for macOS and Linux (x86-64 and aarch64). Recommended, as it is more integrated with your filesystem and overall has less upload and download deltas. See [Nix-based installation](https://librelane.readthedocs.io/en/latest/installation/nix_installation/index.html) in the docs for more info. ### Docker Works for Windows, macOS and Linux (x86-64 and aarch64). See [Docker-based installation](https://librelane.readthedocs.io/en/latest/installation/docker_installation/index.html) in the docs for more info. Do note you'll need to add `--dockerized` right after `librelane` in most CLI invocations. ### Python-only Installation (Advanced, Not Recommended) **You'll need to bring your own compiled utilities**, but otherwise, simply install LibreLane as follows: ```sh python3 -m pip install --upgrade librelane ``` Python-only installations are presently unsupported and entirely at your own risk. ## Usage In the root folder of the repository, you may invoke: ```sh python3 -m librelane --pdk-root <path/to/pdk> </path/to/config.json> ``` To start with, you can try: ```sh python3 -m librelane --pdk-root $HOME/.ciel ./designs/spm/config.json ``` ## Publication If you use LibreLane in your research, please cite the following paper. * M. Shalan and T. Edwards, “Building OpenLANE: A 130nm OpenROAD-based Tapeout-Proven Flow: Invited Paper,” *2020 IEEE/ACM International Conference On Computer Aided Design (ICCAD)*, San Diego, CA, USA, 2020, pp. 1-6. [Paper](https://ieeexplore.ieee.org/document/9256623) ```bibtex @INPROCEEDINGS{9256623, author={Shalan, Mohamed and Edwards, Tim}, booktitle={2020 IEEE/ACM International Conference On Computer Aided Design (ICCAD)}, title={Building OpenLANE: A 130nm OpenROAD-based Tapeout- Proven Flow : Invited Paper}, year={2020}, volume={}, number={}, pages={1-6}, doi={}} ``` ## License and Legal Info LibreLane is a trademark of the [FOSSi Foundation](https://fossi-foundation.org). LibreLane code and binaries are available under [The Apache License, version 2.0](https://www.apache.org/licenses/LICENSE-2.0.txt), except Nix-language files ending with `.nix`, which are available under the [MIT License](https://opensource.org/license/mit) as published by the Open Source Initiative. LibreLane is based on [OpenLane 2](https://github.com/efabless/openlane2) by Efabless Corporation (assets owned by UmbraLogic Technologies LLC): ``` Copyright 2022-2025 UmbraLogic Technologies LLC Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. ``` UmbraLogic Technologies LLC has agreed to relicense all OpenLane 2 Nix code as MIT, for which we are grateful.
text/markdown
null
null
Mohamed Gaber
me@donn.website
null
null
[ "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Programming Language :: Python :: 3.14" ]
[]
null
null
>=3.10
[]
[]
[]
[ "ciel<3,>=2.3.1", "click<8.3,>=8", "cloup<4,>=3.0.5", "deprecated<2,>=1.2.10", "httpx<0.29,>=0.22.0", "klayout<0.32.0,>=0.29.0", "lln-libparse==0.56.*", "lxml>=4.9.0", "psutil>=5.9.0", "pyyaml<7,>=5", "rapidfuzz<4,>=3.9.0", "rich<15,>=12", "semver<4,>=3.0.2", "yamlcore<0.1.0,>=0.0.2" ]
[]
[]
[]
[ "Bug Tracker, https://github.com/librelane/librelane/issues", "Documentation, https://librelane.readthedocs.io", "Homepage, https://librelane.org/", "Repository, https://github.com/librelane/librelane" ]
twine/6.1.0 CPython/3.13.7
2026-02-19T11:50:20.153771
librelane-3.0.0.dev52.tar.gz
301,591
d0/79/804ea3d873ee9bfc8d7c62a1b678a4c15c5482748c69d518f5e24d51cf4f/librelane-3.0.0.dev52.tar.gz
source
sdist
null
false
e11fe673ece8c5f43e8ff9b30c585472
e3de03a3717eced231f708cf3bca73f2a23f5ec55b33900730624089ae0201b7
d079804ea3d873ee9bfc8d7c62a1b678a4c15c5482748c69d518f5e24d51cf4f
Apache-2.0
[]
251
2.4
frequenz-lib-notebooks
0.14.5
Tooling for notebooks.
# Tooling Library for Notebooks [![Build Status](https://github.com/frequenz-floss/frequenz-lib-notebooks/actions/workflows/ci.yaml/badge.svg)](https://github.com/frequenz-floss/frequenz-lib-notebooks/actions/workflows/ci.yaml) [![PyPI Package](https://img.shields.io/pypi/v/frequenz-lib-notebooks)](https://pypi.org/project/frequenz-lib-notebooks/) [![Docs](https://img.shields.io/badge/docs-latest-informational)](https://frequenz-floss.github.io/frequenz-lib-notebooks/) ## Introduction A modular Python toolkit designed to support notebook-based workflows. It provides reusable tools for data ingestion, transformations, visualisation, notifications, and microgrid metadata managers. These tools make the repository ideal for streamlining analytics workflows with minimal setup and building data pipelines, reporting workflows, and alert systems seamlessly in Jupyter or cloud notebooks. ## Supported Platforms The following platforms are officially supported (tested): - **Python:** 3.11 - **Operating System:** Ubuntu Linux 20.04 - **Architectures:** amd64, arm64 ## Contributing If you want to know how to build this project and contribute to it, please check out the [Contributing Guide](CONTRIBUTING.md). ## Quick Start Install the package, open the example notebooks, and explore the available modules. ### Installation ``` # Choose the version you want to install VERSION=0.9.2 pip install frequenz‑lib‑notebooks==$VERSION ``` Then open the prebuilt example notebooks using your preferred interface: - Classic Notebook: `jupyter examples/` - JupyterLab: `jupyter-lab examples/` ⚠️ **Note**: This project does **not** install `jupyter` or `jupyterlab` by default. You will need to install it separately if you want to run notebooks: ``` pip install jupyterlab ``` ### Code Examples #### 📧 Example 1: Generate an Alert Email (HTML Body Only) This example shows how to: - Transform a `pandas` DataFrame of alert records into a structured HTML email using `generate_alert_email`. - Use `AlertEmailConfig` to control layout (e.g., table row limits, sorting by severity). - Integrate microgrid-component alerts cleanly into operational workflows (e.g., for notifications or reporting tools). ``` import pandas as pd from IPython.display import HTML from frequenz.lib.notebooks.alerts.alert_email import ( AlertEmailConfig, generate_alert_email, ) from frequenz.lib.notebooks.notification_utils import format_email_preview # Example alert records dataframe alert_records = pd.DataFrame( [ { "microgrid_id": 1, "component_id": 1, "state_type": "error", "state_value": "UNDERVOLTAGE", "start_time": "2025-03-14 15:06:30", "end_time": "2025-03-14 17:00:00", }, { "microgrid_id": 2, "component_id": 1, "state_type": "state", "state_value": "DISCHARGING", "start_time": "2025-03-14 15:06:30", "end_time": None, }, ] ) # Configuration for email generation alert_email_config = AlertEmailConfig( displayed_rows=10, sort_by_severity=True, ) # Generate the HTML body of the alert email html_email = generate_alert_email( alert_records=alert_records, config=alert_email_config ) # Output the HTML # or send it via email as shown in the next example print(html_email) # or preview it in a nicer format HTML(format_email_preview(subject="Alert Notification", body_html=html_email)) ``` #### 📨 Example 2: Compose and Send Alert Email with Attachments Continuing from Example 1, this snippet builds on the generated HTML email and demonstrates: - Configuring SMTP credentials and recipients. - Attaching both a CSV export of the alert data and optional visual plots. - Sending the email once or scheduling it periodically. Note that the periodic scheduling would make sense when the data also refreshes so as to not send the same email over and over again! ``` import time from datetime import datetime from frequenz.lib.notebooks.alerts.alert_email import ExportOptions, plot_alerts from frequenz.lib.notebooks.notification_service import ( EmailConfig, EmailNotification, SchedulerConfig, ) # Configuration for email notification email_config = EmailConfig( subject="Critical Alert", message=html_email, # Assuming that html_email already exists. See the code example above on how to generate this. recipients=["recipient@example.com"], smtp_server="smtp.example.com", smtp_port=587, smtp_user="user@example.com", smtp_password="password", from_email="alert@example.com", scheduler=SchedulerConfig( send_immediately=True, interval=60, # send every minute duration=3600, # for one hour total ), ) # The SMTP details and sender/recipient details need to be adjusted accordingly # note that the library provides a convenient way to validate the settings via frequenz.lib.notebooks.notification_utils.validate_email_config # Create a notification object email_notification = EmailNotification(config=email_config) # optionally add attachments (a list of files) email_config.attachments = None timestamp = datetime.now().strftime("%Y%m%d_%H%M%S") alert_file_name = f"alert_details_{timestamp}.csv" alert_records.to_csv(alert_file_name, index=False) email_config.attachments = [alert_file_name] # Optionally create a visualisation of the alert records img_path = plot_alerts( records=alert_records, plot_type="all", export_options=ExportOptions( format="png", show=True, ), ) email_config.attachments += img_path if img_path else [] # Send one-off notification email_notification.send() # Or start a periodic scheduler: email_notification.start_scheduler() time.sleep(300) # let it run for 5 minutes email_notification.stop_scheduler() ``` ## Module Overview - **Solar Maintenance App:** Interactive forecasting and visualisation tools tailored to solar installations. - **Notification Service:** Flexible and configurable email dispatching. - **Alert Email Generation:** Embed rich Plotly charts into alert emails, complete with context and summaries. - **Microgrid Configuration:** Manage structured microgrid metadata—location, devices, etc. — consistently across notebooks. For more details about each module/project, please refer to the overview `Wiki` [page](https://github.com/frequenz-floss/frequenz-lib-notebooks/wiki/Frequenz-Lib-Notebooks-%E2%80%90-Overview) which has links to dedicated project pages. The full code documentation can be accessed [here](https://frequenz-floss.github.io/frequenz-lib-notebooks/latest/).
text/markdown
null
Frequenz Energy-as-a-Service GmbH <floss@frequenz.com>
null
null
MIT
frequenz, python, lib, library, frequenz-lib-notebooks, tooling, notebooks
[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3 :: Only", "Topic :: Software Development :: Libraries", "Typing :: Typed" ]
[]
null
null
<4,>=3.11
[]
[]
[]
[ "typing-extensions<5,>=4.14.1", "numpy<3,>=2.3.1", "pandas<3,>=2.3.1", "pyarrow<24.0.0,>=20.0.0", "matplotlib<3.11.0,>=3.9.2", "ipython==9.9.0", "pvlib<0.14.0,>=0.13.0", "python-dotenv<1.3.0,>=0.21.0", "toml<0.11.0,>=0.10.2", "marshmallow-dataclass<9,>=8.7.1", "plotly<6.6.0,>=6.0.0", "kaleido<...
[]
[]
[]
[ "Documentation, https://frequenz-floss.github.io/frequenz-lib-notebooks/", "Changelog, https://github.com/frequenz-floss/frequenz-lib-notebooks/releases", "Issues, https://github.com/frequenz-floss/frequenz-lib-notebooks/issues", "Repository, https://github.com/frequenz-floss/frequenz-lib-notebooks", "Suppo...
twine/6.1.0 CPython/3.13.7
2026-02-19T11:47:44.550898
frequenz_lib_notebooks-0.14.5.tar.gz
2,896,841
ae/a3/a02edf2e2b4da4ed05eee7a5448d03b2c2ca532746f984187d0aff5e1840/frequenz_lib_notebooks-0.14.5.tar.gz
source
sdist
null
false
c010938b9421b30bdc5c7c1af76e6abc
f1f2709267dbbdf4c6cabaf579a59b86ef8d334a33ca23b522361d4407285fd8
aea3a02edf2e2b4da4ed05eee7a5448d03b2c2ca532746f984187d0aff5e1840
null
[ "LICENSE", "LICENSE.txt" ]
285
2.4
pyaedt
0.25.1
High-level Python API for Ansys Electronics Desktop Framework
<!-- --> <a name="readme-top"></a> <!-- *** PyAEDT README --> [![PyAEDT logo](https://raw.githubusercontent.com/ansys/pyaedt/main/doc/source/_static/logo.png)](https://aedt.docs.pyansys.com) <p style="text-align: center;"> <br> English | <a href="README_CN.md">中文</a> </p> [![PyAnsys](https://img.shields.io/badge/Py-Ansys-ffc107.svg?logo=data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAABAAAAAQCAIAAACQkWg2AAABDklEQVQ4jWNgoDfg5mD8vE7q/3bpVyskbW0sMRUwofHD7Dh5OBkZGBgW7/3W2tZpa2tLQEOyOzeEsfumlK2tbVpaGj4N6jIs1lpsDAwMJ278sveMY2BgCA0NFRISwqkhyQ1q/Nyd3zg4OBgYGNjZ2ePi4rB5loGBhZnhxTLJ/9ulv26Q4uVk1NXV/f///////69du4Zdg78lx//t0v+3S88rFISInD59GqIH2esIJ8G9O2/XVwhjzpw5EAam1xkkBJn/bJX+v1365hxxuCAfH9+3b9/+////48cPuNehNsS7cDEzMTAwMMzb+Q2u4dOnT2vWrMHu9ZtzxP9vl/69RVpCkBlZ3N7enoDXBwEAAA+YYitOilMVAAAAAElFTkSuQmCC)](https://docs.pyansys.com/) [![pypi](https://img.shields.io/pypi/v/pyaedt.svg?logo=python&logoColor=white)](https://pypi.org/project/pyaedt/) [![PyPIact](https://static.pepy.tech/badge/pyaedt/month)](https://www.pepy.tech/projects/pyaedt) [![PythonVersion](https://img.shields.io/badge/python-3.10+-blue.svg)](https://www.python.org/downloads/) [![GH-CI](https://github.com/ansys/pyaedt/actions/workflows/ci_cd.yml/badge.svg)](https://github.com/ansys/pyaedt/actions/workflows/ci_cd.yml) [![codecov](https://codecov.io/gh/ansys/pyaedt/branch/main/graph/badge.svg)](https://codecov.io/gh/ansys/pyaedt) [![MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/blog/license/mit) [![black](https://img.shields.io/badge/code%20style-black-000000.svg?style=flat)](https://github.com/psf/black) [![Anaconda](https://anaconda.org/conda-forge/pyaedt/badges/version.svg)](https://anaconda.org/conda-forge/pyaedt) [![pre-commit](https://results.pre-commit.ci/badge/github/ansys/pyaedt/main.svg)](https://results.pre-commit.ci/latest/github/ansys/pyaedt/main) [![deep-wiki](https://deepwiki.com/badge.svg)](https://deepwiki.com/ansys/pyaedt) ## What is PyAEDT? PyAEDT is a Python library that interacts directly with the API for Ansys Electronics Desktop (AEDT) to make scripting simpler. The architecture for PyAEDT can be reused for all AEDT 3D products (HFSS, Icepak, Maxwell 3D, and Q3D Extractor), 2D tools, and Ansys Mechanical inside AEDT. PyAEDT also provides support for circuit tools like Nexxim and system simulation tools like Twin Builder. Finally, PyAEDT provides scripting capabilities in Ansys layout tools like HFSS 3D Layout. The Ansys Electronics Database ([EDB](https://edb.docs.pyansys.com/version/stable/)) is included with PyAEDT as a dependency and is recommended for any automated manipulation and setup of layout data for PCBs, electronic packages, and integrated circuits. The PyAEDT class and method structures simplify operation while reusing information as much as possible across the API. ## Install on CPython from PyPI You can install PyAEDT on CPython 3.8 through 3.12 from PyPI with this command: ```sh pip install pyaedt ``` Install PyAEDT with all extra packages (matplotlib, numpy, pandas, pyvista, ...): ```sh pip install pyaedt[all] ``` > **Note:** The [all] installation target includes `fpdf2` for PDF export functionality, which is under license LGPL v3. For license-sensitive environments, install PyAEDT without [all] and add dependencies individually. You can also install PyAEDT from Conda-Forge with this command: ```sh conda install -c conda-forge pyaedt ``` PyAEDT remains compatible with IronPython and can be still used in the AEDT Framework. ## PyAEDT compatibility requirements PyAEDT has different compatibility requirements based on its version. Below is an overview of the compatibility matrix between PyAEDT, Python versions, and AEDT releases: - PyAEDT Version ≤ 0.8.11: - Python Compatibility: - Compatible with IronPython (Python 2.7). - Compatible with Python 3.7 and versions up to Python 3.11. - AEDT Compatibility: - All tests were conducted using AEDT 2024 R1. - 0.9.0 <= PyAEDT Version < 0.18.0: - Python Compatibility: - Dropped support for python 3.7 and below. - Compatible with Python 3.8 and versions up to Python 3.12. - AEDT Compatibility: - Version 0.9.x has been tested using AEDT 2024 R1. - From versions 0.10.0 to 0.13.3, all tests have been performed with AEDT 2024 R2. - Starting from version 0.14.0, all tests are performed with AEDT 2025 R1. - PyAEDT Version ≥ 0.18.0: - Python Compatibility: - Dropped support for Python 3.8 and 3.9. - Compatible with Python 3.10 and versions up to Python 3.13. - AEDT Compatibility: - All tests were conducted using AEDT 2025 R1. ## About PyAnsys PyAEDT is part of the larger [PyAnsys](https://docs.pyansys.com "PyAnsys") effort to facilitate the use of Ansys technologies directly from Python. PyAEDT is intended to consolidate and extend all existing functionalities around scripting for AEDT to allow reuse of existing code, sharing of best practices, and increased collaboration. ## About AEDT [AEDT](https://www.ansys.com/products/electronics) is a platform that enables true electronics system design. AEDT provides access to the Ansys gold-standard electro-magnetics simulation solutions, such as Ansys HFSS, Ansys Maxwell, Ansys Q3D Extractor, Ansys SIwave, and Ansys Icepak using electrical CAD (ECAD) and Mechanical CAD (MCAD) workflows. In addition, AEDT includes direct links to the complete Ansys portfolio of thermal, fluid, and mechanical solvers for comprehensive multiphysics analysis. Tight integration among these solutions provides unprecedented ease of use for setup and faster resolution of complex simulations for design and optimization. <p align="center"> <img width="100%" src="https://images.ansys.com/is/image/ansys/ansys-electronics-technology-collage?wid=941&op_usm=0.9,1.0,20,0&fit=constrain,0" title="AEDT Applications" herf="https://www.ansys.com/products/electronics" /> </p> `PyAEDT` is licensed under the [MIT License](https://github.com/ansys/pyaedt/blob/main/LICENSE) PyAEDT includes functionality for interacting with the following AEDT tools and Ansys products: - HFSS and HFSS 3D Layout - Icepak - Maxwell 2D, Maxwell 3D, and RMXprt - 2D Extractor and Q3D Extractor - Mechanical - Nexxim - EDB - Twin Builder - EMIT ## Documentation and issues Documentation for the latest stable release of PyAEDT is hosted at [PyAEDT documentation](https://aedt.docs.pyansys.com/version/stable/). In the upper right corner of the documentation's title bar, there is an option for switching from viewing the documentation for the latest stable release to viewing the documentation for the development version or previously released versions. You can also view or download PyAEDT cheat sheets, which are one-page references providing syntax rules and commands for using the PyAEDT API and PyEDB API: - [View](https://cheatsheets.docs.pyansys.com/pyaedt_API_cheat_sheet.png) or [download](https://cheatsheets.docs.pyansys.com/pyaedt_API_cheat_sheet.pdf) the PyAEDT API cheat sheet. - [View](https://cheatsheets.docs.pyansys.com/pyedb_API_cheat_sheet.png) or [download](https://cheatsheets.docs.pyansys.com/pyedb_API_cheat_sheet.pdf) the PyEDB API cheat sheet. On the [PyAEDT Issues](https://github.com/ansys/PyAEDT/issues) page, you can create issues to report bugs and request new features. On the [PyAEDT Discussions](https://github.com/ansys/pyaedt/discussions) page or the [Discussions](https://discuss.ansys.com/) page on the Ansys Developer portal, you can post questions, share ideas, and get community feedback. To reach the project support team, email [pyansys.core@ansys.com](mailto:pyansys.core@ansys.com). ## Dependencies To run PyAEDT, you must have a local licenced copy of AEDT. PyAEDT supports AEDT versions 2022 R1 or newer. ## Contributing For comprehensive information on contributing to the PyAnsys project, see the [PyAnsys developer's guide](https://dev.docs.pyansys.com). Note that PyAEDT uses semantic naming for pull requests (PR). This convention greatly simplifies the review process by providing meaningful information in the PR title. The following [prefixes](https://github.com/ansys/actions/blob/main/check-pr-title/action.yml) should be used for pull request name: - "BUILD" - "CHORE" - "CI" - "DOCS" - "FEAT" - "FIX" - "PERF" - "REFACTOR" - "REVERT" - "STYLE" - "TEST" ## Student version PyAEDT supports AEDT Student versions 2022 R1 and later. For more information, see the [Ansys Electronics Desktop Student - Free Software Download](https://www.ansys.com/academic/students/ansys-electronics-desktop-student) page on the Ansys website. ## Why PyAEDT? A quick and easy approach for automating a simple operation in the AEDT UI is to record and reuse a script. However, disadvantages of this approach are: - Recorded code is dirty and difficult to read and understand. - Recorded scripts are difficult to reuse and adapt. - Complex coding is required by many global users of AEDT. The main advantages of PyAEDT are: - Automatic initialization of all AEDT objects, such as desktop objects like the editor and boundaries - Error management - Log management - Variable management - Compatibility with IronPython and CPython - Simplification of complex API syntax using data objects while maintaining PEP8 compliance. - Code reusability across different solvers - Clear documentation on functions and API - Unit tests of code to increase quality across different AEDT versions ## Example workflow 1. Initialize the ``Desktop`` class with the version of AEDT to use. 2. Initialize the application to use within AEDT. ## Connect to AEDT from Python IDE PyAEDT works both inside AEDT and as a standalone application. This Python library automatically detects whether it is running in an IronPython or CPython environment and initializes AEDT accordingly. PyAEDT also provides advanced error management. Usage examples follow. ## Explicit AEDT declaration and error management ``` python # Launch AEDT 2022 R2 in non-graphical mode from ansys.aedt.core import Desktop, Circuit with Desktop( version="2022.2", non_graphical=False, new_desktop=True, close_on_exit=True, student_version=False, ): circuit = Circuit() ... # Any error here will be caught by Desktop. ... # Desktop is automatically released here. ``` ## Implicit AEDT declaration and error management ``` python # Launch the latest installed version of AEDT in graphical mode from ansys.aedt.core import Circuit with Circuit(version="2022.2", non_graphical=False) as circuit: ... # Any error here will be caught by Desktop. ... # Desktop is automatically released here. ``` ## Remote application call You can make a remote application call on a CPython server or any Windows client machine. On a CPython Server: ``` python # Launch PyAEDT remote server on CPython from ansys.aedt.core.common_rpc import pyaedt_service_manager pyaedt_service_manager() ``` On any Windows client machine: ``` python from ansys.aedt.core.common_rpc import create_session cl1 = create_session("server_name") cl1.aedt(port=50000, non_graphical=False) hfss = Hfss(machine="server_name", port=50000) # your code here ``` ## Variables ``` python from ansys.aedt.core.HFSS import HFSS with HFSS as hfss: hfss["dim"] = "1mm" # design variable hfss["$dim"] = "1mm" # project variable ``` ## Modeler ``` python # Create a box, assign variables, and assign materials. from ansys.aedt.core.hfss import Hfss with Hfss as hfss: hfss.modeler.create_box([0, 0, 0], [10, "dim", 10], "mybox", "aluminum") ``` ## License PyAEDT is licensed under the MIT license. PyAEDT provides an optional PDF export feature via the class `AnsysReport`. It requires the [fpdf2](https://github.com/py-pdf/fpdf2) library, which is licensed under the [GNU Lesser General Public License v3.0 (LGPLv3)](https://www.gnu.org/licenses/lgpl-3.0.en.html#license-text). PyAEDT makes no commercial claim over Ansys whatsoever. This library extends the functionality of AEDT by adding a Python interface to AEDT without changing the core behavior or license of the original software. The use of PyAEDT requires a legally licensed local copy of AEDT. To get a copy of AEDT, see the [Ansys Electronics](https://www.ansys.com/products/electronics) page on the Ansys website. <p style="text-align: right;"> <a href="#readme-top">back to top</a> </p> ## Indices and tables - [Index](https://aedt.docs.pyansys.com/version/stable/genindex.html) - [Module Index](https://aedt.docs.pyansys.com/version/stable/py-modindex.html) - [Search Page](https://aedt.docs.pyansys.com/version/stable/search.html)
text/markdown
null
"ANSYS, Inc." <pyansys.core@ansys.com>
null
"ANSYS, Inc." <pyansys.core@ansys.com>
null
null
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Science/Research", "Intended Audience :: Manufacturing", "Operating System :: OS Independent", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", "Programming Language :: Python ...
[]
null
null
<4,>=3.10
[]
[]
[]
[ "grpcio<1.79,>=1.50.0", "jsonschema", "psutil", "pyedb!=0.28.0,>=0.24.0", "tomli; python_version < \"3.11\"", "tomli-w", "pyyaml", "defusedxml<8.0,>=0.7", "numpy<2.3,>=1.20.0", "pydantic<2.13,>=2.6.4", "ansys-pythonnet>=3.1.0rc3; extra == \"dotnet\"", "cffi<2.1,>=1.16.0; platform_system == \"L...
[]
[]
[]
[ "Bugs, https://github.com/ansys/pyaedt/issues", "Documentation, https://aedt.docs.pyansys.com", "Source, https://github.com/ansys/pyaedt", "Discussions, https://github.com/ansys/pyaedt/discussions", "Releases, https://github.com/ansys/pyaedt/releases", "Changelog, https://github.com/ansys/pyaedt/blob/main...
twine/6.1.0 CPython/3.13.7
2026-02-19T11:47:35.705220
pyaedt-0.25.1.tar.gz
26,878,610
43/a4/ac0e13aa29c2c70261717b3bd290566f6db2eadcf6d594a6001451509c81/pyaedt-0.25.1.tar.gz
source
sdist
null
false
ebdeb3dc6843aa56c684e168710516ec
dea0fba522525ebf82d467b1ca9126bc9d59811ee95e7d0d79fd0ab8699f3044
43a4ac0e13aa29c2c70261717b3bd290566f6db2eadcf6d594a6001451509c81
MIT
[ "LICENSE" ]
1,714
2.4
dod-outline-discovery
0.1.2
Document Outline Discovery.
# DoD (Document Outline Discovery) > This project is intended to make local deployment and usage of PageIndex easier. It is not an official PageIndex service. Check their official PageIndex repo [here](https://github.com/VectifyAI/PageIndex). DoD is a local-first document structure extraction toolkit built around [PageIndex](https://pageindex.ai/). What it does: - ingests PDFs and builds normalized page artifacts - extracts page-level text (`pymupdf` or `pytesseract`) - generates hierarchical TOC/document outline trees - exposes both **CLI** and **server+MCP** interfaces for agent workflows Use cases: - document-grounded Q&A assistants over private document libraries - TOC/outline extraction for scanned or OCR-heavy PDFs - page-targeted retrieval pipelines for downstream RAG/agent systems Advantages: - local deployment for sensitive documents - structured outputs (TOC tree + page tables + page images) - agent-friendly retrieval via stable server and MCP tools Why this can make more sense than traditional RAG for long manuals: - instead of flat chunk retrieval, PageIndex builds explicit document structure (sections/subsections + page mapping) - this improves navigation, targeted retrieval, and answer grounding for long technical documents - see PageIndex's technical-manual discussion: https://pageindex.ai/blog/technical-manuals Example (Codex using DoD MCP for section-grounded Q&A): ![Codex + DoD MCP example](mcp_example_pics/mcp-example-codex.png) --- ## Table of Contents - [Project Structure](#project-structure) - [0. LLM API Configuration Required](#0-llm-api-configuration-required) - [1. Install and Dev](#1-install-and-dev) - [1.1 Use as a PyPI package](#11-use-as-a-pypi-package) - [2. Run As A Package CLI](#2-run-as-a-package-cli) - [3. Run As A Server](#3-run-as-a-server) - [3.1 Start server](#31-start-server) - [3.2 Health check](#32-health-check) - [3.3 Make requests](#33-make-requests) - [4. Output What And Where](#4-output-what-and-where) - [4.3 Configure output paths PyPI package](#43-configure-output-paths-pypi-package) - [5. Request Fields Server `/v1/digest`](#5-request-fields-server-v1digest) - [6. Simple MCP Setup](#6-simple-mcp-setup) - [6.1 Install from PyPI for MCP use](#61-install-from-pypi-for-mcp-use) - [6.2 Start DoD HTTP server](#62-start-dod-http-server) - [6.3 Configure MCP client recommended](#63-configure-mcp-client-recommended) - [6.4 Optional manual MCP run debug only](#64-optional-manual-mcp-run-debug-only) - [6.5 Available MCP tools](#65-available-mcp-tools) - [6.6 Example agent skill](#66-example-agent-skill) - [7. Third-Party Licensing](#7-third-party-licensing) DoD turns a document into: 1. `page_table.jsonl` (page text + metadata) 2. `toc_tree.json` (hierarchical Table-of-Contents tree) 3. `image_page_table.jsonl` (page-image index, including image paths and image payload fields) 4. `images/` (all pages rendered as image files) Core flow: 1. PDF normalization to page images 2. Text extraction (`pymupdf` / `pytesseract`) 3. TOC generation (PageIndex) 4. Artifact writing (JSON/JSONL + manifest) ## Project Structure - `src/DoD/cli/` - CLI entrypoint - `src/DoD/pipeline.py` - end-to-end pipeline orchestration - `src/DoD/normalize/` - PDF/image normalization to per-page images - `src/DoD/text_extractor/` - text extraction backends - `src/DoD/page_table.py` - page table data model + writer - `src/DoD/pageindex/` - PageIndex TOC builder - `src/DoD/toc/` - TOC adapters - `src/DoD/server/` - FastAPI server mode - `src/scripts/` - executable entrypoints (`dod`, `dod-server`, `dod-mcp`) - `.agents/skills/` - example agent skills - `makefile` - install/check/test developer workflow - `src/DoD/conf/config.yaml` - default configuration ## 0. LLM API Configuration Required Before running either the CLI package mode or the server mode, set an OpenAI-compatible endpoint and API key: ```bash export PAGEINDEX_API_KEY="<your_api_key>" export PAGEINDEX_BASE_URL="<your_openai_compatible_base_url>" ``` Example (Snowflake Cortex): ```bash export PAGEINDEX_API_KEY="<snowflake_pat>" export PAGEINDEX_BASE_URL="https://<account-identifier>.snowflakecomputing.com/api/v2/cortex/v1" ``` Then choose any model available on your configured endpoint via `toc.model`. ## 1. Install and Dev Use these commands for setup and day-to-day development: ```bash make install # bootstrap toolchain + Python deps + hooks make check # lint + format + type-check make test # run test suite ``` ### 1.1 Use as a PyPI package Install from PyPI: ```bash pip install "dod-outline-discovery[pageindex,text_extractor,pdf,server,mcp]" ``` Then run: ```bash dod --help dod-server --help dod-mcp --help ``` One-off execution without a persistent install: ```bash uvx --from "dod-outline-discovery[pageindex,text_extractor,pdf,server,mcp]" dod --help uvx --from "dod-outline-discovery[pageindex,text_extractor,pdf,server,mcp]" dod-server --help uvx --from "dod-outline-discovery[pageindex,text_extractor,pdf,server,mcp]" dod-mcp --help ``` ## 2. Run As A Package CLI Run one document: Choose extractor first: - use `text_extractor.backend=pymupdf` for PDFs with built-in text layers - use `text_extractor.backend=pytesseract` for image-based/scanned PDFs ```bash dod \ input_path=/path/to/document.pdf \ text_extractor.backend=pymupdf \ toc.backend=pageindex \ toc.model=claude-sonnet-4-5 ``` If you are running from source repo instead of a PyPI install, use `uv run dod`. Where output is written: - Hydra run dir: `outputs/<YYYY-MM-DD>/<HH-MM-SS>/` - Artifacts folder: `outputs/<YYYY-MM-DD>/<HH-MM-SS>/artifacts/` Main artifact files: - `page_table.jsonl` - `image_page_table.jsonl` - `toc_tree.json` - `manifest.json` ## 3. Run As A Server ### 3.1 Start server ```bash export DOD_SERVER_HOST=0.0.0.0 export DOD_SERVER_PORT=8000 export DOD_SERVER_MAX_CONCURRENT_DOCS=4 export DOD_SERVER_JOB_TIMEOUT_SECONDS=300 export DOD_SERVER_WORK_DIR=outputs/server_jobs dod-server ``` If you are running from source repo instead of a PyPI install, use `uv run dod-server`. ### 3.2 Health check ```bash curl http://localhost:8000/healthz ``` ### 3.3 Make requests #### 3.3.1 Single PDF wait for final result ```bash curl -s -X POST "http://localhost:8000/v1/digest" \ -F "file=@/path/to/document.pdf" \ -F "text_extractor_backend=pymupdf" \ -F "toc_backend=pageindex" \ -F "toc_model=claude-sonnet-4-5" \ -F "toc_concurrent_requests=4" \ > result.json ``` This call blocks until the job is done and writes full result JSON to `result.json`. #### 3.3.2 Async job submit then poll Submit: ```bash SUBMIT_JSON=$(curl -s -X POST "http://localhost:8000/v1/digest?wait=false" \ -F "file=@/path/to/document.pdf" \ -F "text_extractor_backend=pymupdf" \ -F "toc_backend=pageindex" \ -F "toc_model=claude-sonnet-4-5") JOB_ID=$(echo "$SUBMIT_JSON" | jq -r '.job_id') JOB_REF=$(echo "$SUBMIT_JSON" | jq -r '.job_ref') ``` Check status: ```bash curl -s "http://localhost:8000/v1/jobs/$JOB_REF" ``` Get final result: ```bash curl -s "http://localhost:8000/v1/jobs/$JOB_REF/result" > result.json ``` #### 3.3.3 Process More Than One PDF At The Same Time Use async submit (`wait=false`) + parallel curl: ```bash mkdir -p jobs printf "%s\n" \ "/path/to/a.pdf" \ "/path/to/b.pdf" \ "/path/to/c.pdf" \ | xargs -I{} -P 3 sh -c ' name=$(basename "{}" .pdf) curl -s -X POST "http://localhost:8000/v1/digest?wait=false" \ -F "file=@{}" \ -F "text_extractor_backend=pymupdf" \ -F "toc_backend=pageindex" \ -F "toc_model=claude-sonnet-4-5" \ -F "toc_concurrent_requests=4" \ > "jobs/${name}.submit.json" ' ``` Poll and download all results: ```bash for f in jobs/*.submit.json; do job_id=$(jq -r '.job_id' "$f") name=$(basename "$f" .submit.json) until curl -sf "http://localhost:8000/v1/jobs/$job_id/result" > "jobs/${name}.result.json"; do sleep 2 done done ``` Notes: - `-P 3` controls how many submit requests run in parallel. - Server-side processing concurrency is capped by `DOD_SERVER_MAX_CONCURRENT_DOCS`. - TOC runs in strict mode: if PageIndex fails, the job status becomes `failed` (no fallback TOC). ## 4. Output What And Where ### 4.1 Output in API response JSON `/v1/digest` (sync) or `/v1/jobs/{job_id}/result` returns: - `result.toc_tree` - TOC tree JSON - `result.page_table` - page records (JSON array parsed from JSONL) - `result.image_page_table` - image records (JSON array parsed from JSONL) - `result.artifact_paths` - filesystem paths for written artifacts - `result.manifest` - run metadata + config Convert arrays back to JSONL if needed: ```bash jq -c '.result.page_table[]' result.json > page_table.jsonl jq -c '.result.image_page_table[]' result.json > image_page_table.jsonl jq '.result.toc_tree' result.json > toc_tree.json ``` ### 4.2 Output on disk Server mode For each job: - Job folder: `${DOD_SERVER_WORK_DIR}/<job_id>/` - Input copy: `${DOD_SERVER_WORK_DIR}/<job_id>/input.pdf` - Artifact folder: `${DOD_SERVER_WORK_DIR}/<job_id>/artifacts/` Inside `artifacts/`: - `page_table.jsonl` - `image_page_table.jsonl` - `toc_tree.json` - `manifest.json` Server-level job index: - `${DOD_SERVER_WORK_DIR}/jobs.json` - persists job metadata across server restarts - used by `GET /v1/jobs` and MCP `list_jobs()` Timeout: - default from `src/DoD/conf/config.yaml` → `server.job_timeout_seconds` - runtime override via `DOD_SERVER_JOB_TIMEOUT_SECONDS` ### 4.3 Configure output paths PyPI package When installed from PyPI, outputs are still filesystem-based and default to the current working directory. - CLI (`dod`): defaults to Hydra output under `outputs/<YYYY-MM-DD>/<HH-MM-SS>/artifacts` - Server (`dod-server`): defaults to `outputs/server_jobs` Use absolute paths in production/local deployments: ```bash export DOD_SERVER_WORK_DIR="/absolute/path/to/server_jobs" export DOD_LLM_CACHE_DIR="/absolute/path/to/llm_cache" ``` Optional CLI override for one run: ```bash dod input_path=/path/to/document.pdf artifacts.output_dir=/absolute/path/to/artifacts ``` ## 5. Request Fields Server `/v1/digest` Multipart form fields: - `file` (required, `.pdf`) - `text_extractor_backend` (optional) - use `pymupdf` for PDFs with built-in text layers - use `pytesseract` for image-based/scanned PDFs - `normalize_max_pages` (optional int) - `toc_backend` (optional, typically `pageindex`) - `toc_model` (optional model name) - `toc_concurrent_requests` (optional int) - `toc_check_page_num` (optional int) - `toc_api_key` (optional per-request override) - `toc_api_base_url` (optional per-request override) Query parameter: - `wait` (default `true`) - `true`: request returns when job finishes - `false`: request returns immediately with job metadata (`job_id`, `job_ref`, status/result URLs) ## 6. Simple MCP Setup If you want agents to call DoD as tools, use the included MCP wrapper at `src/scripts/dod_mcp.py`. Agents do not upload PDFs. Human users submit jobs first via `/v1/digest`, then agents use `job_ref` to retrieve targeted outputs. `make install` already installs all extras including MCP/server dependencies. ### 6.1 Install from PyPI for MCP use ```bash uv tool install "dod-outline-discovery[pageindex,text_extractor,pdf,server,mcp]" ``` This installs `dod`, `dod-server`, and `dod-mcp` as system tools. ### 6.2 Start DoD HTTP server ```bash dod-server ``` If you are running from source repo instead of PyPI install, use `uv run dod-server`. ### 6.3 Configure MCP client recommended Most MCP hosts should launch `dod-mcp` themselves via command config, for example (`~/.codex/config.toml`): ```toml [mcp_servers.dod_mcp] command = "dod-mcp" ``` Alternative without tool install: ```toml [mcp_servers.dod_mcp] command = "uvx" args = ["--from", "dod-outline-discovery[pageindex,text_extractor,pdf,server,mcp]", "dod-mcp"] ``` In this mode, do not start `dod-mcp` manually. Keep only `dod-server` running. ### 6.4 Optional manual MCP run debug only ```bash dod-mcp ``` If you are running from source repo instead of a PyPI install, use `uv run dod-mcp`. Use this only for debugging MCP transport behavior. ### 6.5 Available MCP tools - `list_jobs()` - Returns: `{ jobs: [{ file_name, job_id, job_ref, status, created_at }] }`. - `get_toc(job_ref)` - Returns: `{ job_id, job_ref, status, toc_tree }`. - `get_page_texts(job_ref, pages)` - Returns: `{ job_id, job_ref, status, requested_pages, returned_pages, pages }` where `pages` is a selected subset of `{ page_id, text }`. - `get_page_images(job_ref, pages, mode)` - Returns: `{ job_id, job_ref, status, mode, requested_pages, returned_pages, pages }` where `pages` is a selected subset of `{ page_id, image_path }` (`mode=path`) or `{ page_id, image_b64 }` (`mode=base64`). - `pages` accepts flexible specs like `"110,111,89-100"`. - Retrieval guardrails come from `src/DoD/conf/config.yaml` under `retrieval`: - `max_chars_per_page` (`null` means full page text) - `max_pages_per_call` ### 6.6 Example agent skill An example Codex skill for DoD library Q&A is included at: - `.agents/skills/dod-library-qa/SKILL.md` ## 7. Third-Party Licensing This project includes third-party license attributions in: - `THIRD_PARTY_NOTICES.md`
text/markdown
null
Tao Tang <ttan@habitus.dk>
null
null
null
null
[]
[]
null
null
>=3.11
[]
[]
[]
[ "hydra-core>=1.3.2", "omegaconf>=2.3.0", "tqdm>=4.67.2", "httpx>=0.28.1; extra == \"mcp\"", "mcp>=1.12.4; extra == \"mcp\"", "openai>=1.0.0; extra == \"pageindex\"", "pymupdf>=1.24.0; extra == \"pageindex\"", "pypdf2>=3.0.1; extra == \"pageindex\"", "tiktoken>=0.7.0; extra == \"pageindex\"", "pdf2...
[]
[]
[]
[]
twine/6.2.0 CPython/3.11.11
2026-02-19T11:47:29.974692
dod_outline_discovery-0.1.2.tar.gz
463,024
46/24/ca1cdff4b114973772f6e76421dc07a3d83588fae8fd80bf50c21b369684/dod_outline_discovery-0.1.2.tar.gz
source
sdist
null
false
fa7c130f254fe134b0c9e3cd820fe803
c1fd59eb32d2060e53826ccdda550994297a26ef604a78e41b3849564c80291e
4624ca1cdff4b114973772f6e76421dc07a3d83588fae8fd80bf50c21b369684
MIT
[ "LICENSE" ]
238
2.4
docalign
0.1.1
Auto-fix alignment in markdown docs
# Overview CLI utility that auto-fixes alignment issues in markdown documentation files - tables, box-drawing diagrams, list descriptions, and more. ``` ┌───────────────────┐ ┌───────────────────┐ │ ┌────┐ ┌────┐ │ │ ┌────┐ ┌────┐ │ │ │ A │ │ B │ │ │ │ A │ │ B │ │ │ └──┬─┘ └──┬─┘ │ │ └──┬─┘ └──┬─┘ │ │ │ │ │ │ │ │ │ │ └────┬───┘ │ --> │ └────┬───┘ │ │ v │ │ v │ │ ┌──────┐ │ │ ┌──────┐ │ │ │ C │ │ │ │ C │ │ │ └──────┘ │ │ └──────┘ │ └───────────────────┘ └───────────────────┘ ``` <div align="center"> <details> <summary>See all 12 checks</summary> <div align="left"> ### [Tables](tests/fixtures/checks/tables) Aligns columns and ensures `| content |` spacing in cells. ``` | Service | Usage | | Service | Usage | |----------------|-------------------------------| |------------|------------------------------| | Linear API | Status transitions, comments| --> | Linear API | Status transitions, comments | | GitHub API| Repo clone, PR creation | | GitHub API | Repo clone, PR creation | ``` ### [List descriptions](tests/fixtures/checks/list-descs) Aligns the separator dash in list item descriptions. ``` - docs/repo.md - mirrors CI steps - docs/repo.md - mirrors CI steps - docs/guides/testing-strategy.md - test suite --> - docs/guides/testing-strategy.md - test suite ``` ### [Definition lists](tests/fixtures/checks/def-lists) Aligns the `:` separator in key-value list items. ``` - timeout: 30s - timeout: 30s - retries: 3 --> - retries: 3 - max-connections: 100 - max-connections: 100 ``` ### [Box widths](tests/fixtures/checks/box-widths) Ensures all lines in a box group have the same total length. ``` ┌──────────────┐ ┌───────────────┐ │ Linear UI │ --> │ Linear UI │ │ (userscript)│ │ (userscript) │ └──────────────┘ └───────────────┘ ``` ### [Rail alignment](tests/fixtures/checks/rails) Aligns vertically adjacent box chars to the same column. ``` ┌──────┬──────┐ ┌──────┬──────┐ │ │ │ │ │ │ │ │ │ --> │ │ │ │ │ │ │ │ │ └──────┴──────┘ └──────┴──────┘ ``` ### [Arrow alignment](tests/fixtures/checks/arrows) Aligns standalone `v`/`^` arrows with their source pipes. ``` ┌──────┐ ┌──────┐ │ step │ │ step │ └──┬───┘ └──┬───┘ │ --> │ v v ┌──────┐ ┌──────┐ │ next │ │ next │ └──────┘ └──────┘ ``` ### [Pipe continuity](tests/fixtures/checks/pipes) Traces from T-junctions to detect drifted connector pipes. ``` ┌──────┬──────┐ ┌──────┬──────┐ │ src │ dest │ │ src │ dest │ └──────┴──┬───┘ └──────┬──────┘ │ --> │ │ │ │ │ ┌──────┴──────┐ ┌──────┴──────┐ │ output │ │ output │ └─────────────┘ └─────────────┘ ``` ### [Box spacing](tests/fixtures/checks/box-spacing) Ensures minimum right-side spacing between content and box wall. ``` ┌───────────┐ ┌────────────┐ │ errors[] │ │ errors[] │ │ (strings)│ --> │ (strings) │ └───────────┘ └────────────┘ ``` ### [Box walls](tests/fixtures/checks/box-walls) Verifies nested box right walls match their opening/closing borders. ``` ┌──────────────────┐ ┌──────────────────┐ │ content here │ │ content here │ │ more text │ --> │ more text │ └────────────────┘ └──────────────────┘ ``` ### [Box padding](tests/fixtures/checks/box-padding) Normalizes left-padding of content lines inside boxes. ``` ┌──────────────┐ ┌──────────────┐ │ validateAuth │ │ validateAuth │ │ compare with │ --> │ compare with │ └──────────────┘ └──────────────┘ ``` ### [Horizontal arrows](tests/fixtures/checks/horiz-arrows) Closes gaps between arrow tips and box walls. ``` ┌──────┐ ┌──────┐ ┌──────┐ ┌──────┐ │ Src │─────────>│ Dest │ --> │ Src │─────────>│ Dest │ └──────┘ └──────┘ └──────┘ └──────┘ ``` ### [Wide chars](tests/fixtures/checks/wide-chars) Detects ambiguous/double-width Unicode chars in code blocks. ``` ┌──────────┐ │ ▶ start │ --> warning: wide char '▶' (U+25B6) at col 2 │ ● status │ └──────────┘ ``` </div> </details> </div> ## Motivation In the era of agentic engineering, documentation is the most critical artifact in any codebase. It guides both humans and AI agents. When docs are visually harmonious - with aligned columns, consistent box widths, and straight connector lines - they become easier to read, parse, and maintain by everyone. ## Features - 3 modes - check (default), auto-fix in place, or unified diff - flexible paths - files, directories, or glob patterns (e.g. `"docs/**/*.md"`) - CI-friendly - exit code 0 when aligned, 1 when issues found - 12 checks - tables, boxes, arrows, pipes, lists (see examples above) ## Commands ```bash docalign <path> # check-only (default) - detect issues, no writes docalign --check <path> # explicit check-only (same as default) docalign --fix <path> # auto-fix files in place docalign --diff <path> # show unified diff of what would change docalign --verbose <path> # show actionable hints with each error docalign --ignore tables,pipes <path> # skip specific checks docalign --help # show help docalign --version # show version ``` Paths can be files, directories, or glob patterns (e.g. `"docs/**/*.md"`). Check names for `--ignore`: tables, box-widths, box-padding, box-spacing, horiz-arrows, box-walls, rails, arrows, pipes, list-descs, def-lists, wide-chars. Exit codes: 0 = all aligned, 1 = issues found. ### Install ```bash pipx install docalign # pip install docalign ``` ### Update ```bash pipx upgrade docalign # pip install --upgrade docalign ``` ### Uninstall ```bash pipx uninstall docalign # pip uninstall docalign ``` ## Development ```bash python3 -m venv .venv source .venv/bin/activate pip install -e ".[dev]" pytest -v ``` ```bash # dev alias (mdalign) ln -s $(pwd)/.venv/bin/docalign ~/.local/bin/docalignd # install rm ~/.local/bin/docalignd # remove ```
text/markdown
null
null
null
null
null
null
[]
[]
null
null
>=3.9
[]
[]
[]
[ "bump2version>=1; extra == \"dev\"", "pytest>=7; extra == \"dev\"", "ruff>=0.9; extra == \"dev\"", "towncrier>=23; extra == \"dev\"" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-19T11:46:55.023075
docalign-0.1.1.tar.gz
49,997
aa/ed/e21bef858d2312bd0b80661ca2ce62059a182196c6923af9707f994b6d9b/docalign-0.1.1.tar.gz
source
sdist
null
false
fd9eff6363325e3fb2da6fe2edf19838
4a0dc0213b4981b8db18eff51ac73f566cf5716dca511b55defb143d960e9a03
aaede21bef858d2312bd0b80661ca2ce62059a182196c6923af9707f994b6d9b
null
[ "LICENSE" ]
292
2.4
dmsc
0.1.8b2
Dunimd Middleware Service - A high-performance Rust middleware framework with modular architecture
<div align="center"> <h1 style="display: flex; flex-direction: column; align-items: center; gap: 8px; margin-bottom: 8px;"> <span style="display: flex; align-items: center; gap: 12px;"><img src="assets/svg/dmsc.svg" width="36" height="36" alt="DMSC">Dunimd Middleware Service</span> </h1> English | [简体中文](README.zh.md) [Help Documentation](doc/en/index.md) | [Changelog](CHANGELOG.md) | [Security](SECURITY.md) | [Contributing](CONTRIBUTING.md) | [Code of Conduct](CODE_OF_CONDUCT.md) <a href="https://space.bilibili.com/3493284091529457" target="_blank"> <img alt="BiliBili" src="https://img.shields.io/badge/BiliBili-Dunimd-00A1D6?style=flat-square&logo=bilibili"/> </a> <a href="https://x.com/Dunimd2025" target="_blank"> <img alt="X" src="https://img.shields.io/badge/X-Dunimd-000000?style=flat-square&logo=x"/> </a> <a href="https://gitee.com/dunimd" target="_blank"> <img alt="Gitee" src="https://img.shields.io/badge/Gitee-Dunimd-C71D23?style=flat-square&logo=gitee"/> </a> <a href="https://github.com/mf2023/DMSC" target="_blank"> <img alt="GitHub" src="https://img.shields.io/badge/GitHub-DMSC-181717?style=flat-square&logo=github"/> </a> <a href="https://huggingface.co/dunimd" target="_blank"> <img alt="Hugging Face" src="https://img.shields.io/badge/Hugging%20Face-Dunimd-FFD21E?style=flat-square&logo=huggingface"/> </a> <a href="https://modelscope.cn/organization/dunimd" target="_blank"> <img alt="ModelScope" src="https://img.shields.io/badge/ModelScope-Dunimd-1E6CFF?style=flat-square&logo=data:image/svg+xml;base64,PHN2ZyB3aWR0aD0iMTQiIGhlaWdodD0iMTQiIHZpZXdCb3g9IjAgMCAxNCAxNCIgZmlsbD0ibm9uZSIgeG1sbnM9Imh0dHA6Ly93d3cudzMub3JnLzIwMDAvc3ZnIj4KPHBhdGggZD0iTTcuMDA2IDBDMy4xNDIgMCAwIDMuMTQyIDAgNy4wMDZTMy4xNDIgMTQuMDEyIDcuMDA2IDE0LjAxMkMxMC44NyAxNC4wMTIgMTQuMDEyIDEwLjg3IDE0LjAxMiA3LjAwNkMxNC4wMTIgMy4xNDIgMTAuODcgMCA3LjAwNiAwWiIgZmlsbD0iIzFFNkNGRiIvPgo8L3N2Zz4K"/> </a> <a href="https://crates.io/crates/dmsc" target="_blank"> <img alt="Crates.io" src="https://img.shields.io/badge/Crates-DMSC-000000?style=flat-square&logo=rust"/> </a> <a href="https://pypi.org/project/dmsc/" target="_blank"> <img alt="PyPI" src="https://img.shields.io/badge/PyPI-DMSC-3775A9?style=flat-square&logo=pypi"/> </a> <a href="https://docs.rs/dmsc/latest/dmsc/c/index.html" target="_blank"> <img alt="C API" src="https://img.shields.io/badge/C%2FC%2B%2B-API-00599C?style=flat-square&logo=c"/> </a> **DMSC (Dunimd Middleware Service)** — A high-performance Rust middleware framework that unifies backend infrastructure. Built for enterprise-scale with modular architecture, built-in observability, and distributed systems support. </div> <h2 align="center">🏗️ Core Architecture</h2> ### 📐 Modular Design DMSC adopts a highly modular architecture with 18 core modules, enabling on-demand composition and seamless extension: <div align="center"> | Module | Description | |:--------|:-------------| | **auth** | Authentication & authorization (JWT, OAuth, permissions) | | **cache** | Multi-backend cache abstraction (Memory, Redis, Hybrid) | | **config** | Multi-source configuration management with hot reload | | **core** | Runtime, error handling, and service context | | **database** | Database abstraction with PostgreSQL, MySQL, SQLite support | | **device** | Device control, discovery, and intelligent scheduling | | **fs** | Secure file system operations and management | | **gateway** | API gateway with load balancing, rate limiting, and circuit breaking | | **grpc** | gRPC server and client support with Python bindings (requires `grpc` feature) | | **hooks** | Lifecycle event hooks (Startup, Shutdown, etc.) | | **log** | Structured logging with tracing context integration | | **module_rpc** | Inter-module RPC communication for distributed method calls | | **observability** | Metrics, tracing, and Grafana integration | | **database.orm** | Type-safe ORM with repository pattern, query builder, and Python bindings | | **protocol** | Protocol abstraction layer for multi-protocol support (requires `pyo3` feature) | | **queue** | Distributed queue abstraction (Kafka, RabbitMQ, Redis, Memory) | | **service_mesh** | Service discovery, health checking, and traffic management | | **validation** | Input validation and data sanitization utilities | | **ws** | WebSocket server support with Python bindings (requires `websocket` feature) | | **c** | C/C++ FFI bindings for cross-language integration (requires `c` feature) | </div> > **Note**: Some modules require specific feature flags: > - `grpc`: gRPC support (`--features grpc`) > - `websocket`: WebSocket support (`--features websocket`) > - `protocol`: Protocol abstraction layer (`--features protocol` or `full`) > - `c`: C/C++ FFI bindings (`--features c`) ### 🚀 Key Features #### 🔍 Distributed Tracing - W3C Trace Context standard implementation - Full-chain TraceID/SpanID propagation - Baggage data transmission for business context - Multi-language compatibility (Java, Go, Python) - Automatic span creation via `#[tracing::instrument]` attribute #### 📊 Enterprise Observability - Native Prometheus metrics export - Counter, Gauge, Histogram, Summary metric types - Out-of-the-box Grafana dashboard integration - Real-time performance statistics with quantile calculation - Full-stack metrics (CPU, memory, I/O, network) #### 🤖 Intelligent Device Management - Auto-discovery and registration - Efficient resource pool management - Policy-based scheduling with priority support - Dynamic load balancing - Complete device lifecycle management #### 📝 Structured Logging - JSON and text format support - Configurable sampling rates - Intelligent log rotation - Automatic tracing context inclusion - DEBUG/INFO/WARN/ERROR log levels #### ⚙️ Flexible Configuration - Multi-source loading (files, environment variables, runtime) - Hot configuration updates - Modular architecture for on-demand composition - Plugin-based extension mechanism #### 📁 Secure File System - Unified project root directory management - Atomic file operations - Categorized directory structure - JSON data persistence - Secure path handling <h2 align="center">🛠️ Installation & Environment</h2> ### Prerequisites - **Rust**: 1.65+ (2021 Edition) - **Cargo**: 1.65+ - **Platforms**: Linux, macOS, Windows ### Build Dependencies Some features require additional system dependencies: | Dependency | Required For | Installation | |:-----------|:-------------|:-------------| | **protoc** | etcd feature (Protocol Buffers) | [Protocol Buffers](https://protobuf.dev/downloads/) | | **CMake + C++ compiler** | kafka feature (rdkafka) | See instructions below | #### Installing protoc **Windows:** ```powershell # Using chocolatey choco install protoc # Or download from GitHub releases # https://github.com/protocolbuffers/protobuf/releases ``` **macOS:** ```bash brew install protobuf ``` **Linux (Ubuntu/Debian):** ```bash sudo apt-get update sudo apt-get install -y protobuf-compiler ``` **Linux (CentOS/RHEL):** ```bash sudo yum install -y protobuf-compiler ``` #### Installing CMake and C++ compiler (for Kafka support) **Windows:** ```powershell # CMake is usually installed with Visual Studio # Or download from: https://cmake.org/download/ # Using chocolatey choco install cmake ``` **macOS:** ```bash # CMake and C++ compiler (Xcode Command Line Tools) xcode-select --install # Or using Homebrew brew install cmake ``` **Linux (Ubuntu/Debian):** ```bash sudo apt-get update sudo apt-get install -y cmake build-essential ``` **Linux (CentOS/RHEL):** ```bash sudo yum install -y cmake gcc-c++ make ``` ### Quick Setup Add DMSC to your project's `Cargo.toml`: ```toml [dependencies] dmsc = { git = "https://github.com/mf2023/DMSC" } ``` Or use cargo add: ```bash cargo add dmsc --git https://github.com/mf2023/DMSC ``` <h2 align="center">⚡ Quick Start</h2> ### Core API Usage ```rust use dmsc::prelude::*; #[tokio::main] async fn main() -> DMSCResult<()> { // Build service runtime let app = DMSCAppBuilder::new() .with_config("config.yaml")? .with_logging(DMSCLogConfig::default()) .with_observability(DMSCObservabilityConfig::default()) .build()?; // Run business logic app.run(|ctx: &DMSCServiceContext| async move { ctx.logger().info("service", "DMSC service started")?; // Your business code here Ok(()) }).await } ``` ### Observability Example ```rust use dmsc::prelude::*; use dmsc::observability::{DMSCTracer, DMSCSpanKind, DMSCSpanStatus}; #[tracing::instrument(name = "user_service", skip(ctx))] async fn get_user(ctx: &DMSCServiceContext, user_id: u64) -> DMSCResult<User> { let user = fetch_user_from_db(user_id).await?; Ok(user) } ``` Or using DMSCTracer directly: ```rust use dmsc::prelude::*; use dmsc::observability::DMSCTracer; async fn get_user(ctx: &DMSCServiceContext, user_id: u64) -> DMSCResult<User> { let tracer = DMSCTracer::new(1.0); let _span = tracer.span("get_user") .with_attribute("user_id", user_id.to_string()) .start(); let user = fetch_user_from_db(user_id).await?; Ok(user) } ``` <h2 align="center">🔧 Configuration</h2> ### Configuration Example ```yaml # config.yaml service: name: "my-service" version: "1.0.0" logging: level: "info" file_format: "json" file_enabled: true console_enabled: true observability: metrics_enabled: true tracing_enabled: true prometheus_port: 9090 resource: providers: ["cpu", "gpu", "memory"] scheduling_policy: "priority_based" ``` ### Configuration Sources DMSC supports multiple configuration sources in order of priority (lowest to highest): 1. Configuration files (YAML, TOML, JSON) 2. Custom configuration via code 3. Environment variables (prefixed with `DMSC_`) <h2 align="center">🧪 Development & Testing</h2> ### Running Tests ```bash # Run all tests cargo test # Run specific test module cargo test cache # Run with verbose output cargo test -- --nocapture ``` <h2 align="center">❓ Frequently Asked Questions</h2> **Q: How to add a new module?** A: Implement the `DMSCModule` trait and register it via `DMSCAppBuilder::with_module`. **Q: How to configure logging level?** A: Set `logging.level` in the configuration file, supporting DEBUG/INFO/WARN/ERROR levels. **Q: How to enable metrics export?** A: Set `observability.metrics_enabled: true` and configure `prometheus_port` in the configuration file. **Q: How to extend configuration sources?** A: Implement a custom configuration loader and register it with `DMSCConfigManager`. **Q: How to handle asynchronous tasks?** A: Use `DMSCAppBuilder::with_async_module` to add async modules, the framework handles async lifecycle automatically. <h2 align="center">🌏 Community & Citation</h2> - Welcome to submit Issues and PRs! - Gitee: https://gitee.com/dunimd/dmsc.git - Github: https://github.com/mf2023/DMSC.git <div align="center"> ## 📄 License & Open Source Agreements ### 🏛️ Project License <p align="center"> <a href="LICENSE"> <img src="https://img.shields.io/badge/License-Apache%202.0-blue.svg" alt="Apache License 2.0"> </a> </p> This project uses **Apache License 2.0** open source agreement, see [LICENSE](LICENSE) file. ### 📋 Dependency Package Open Source Agreements Open source packages and their agreement information used by this project: ### Dependencies License <div align="center"> | 📦 Package | 📜 License | |:-----------|:-----------| | serde | Apache 2.0 | | serde_json | MIT | | serde_yaml | MIT | | tokio | MIT | | prometheus | Apache 2.0 | | redis | MIT | | hyper | MIT | | lapin | Apache 2.0 | | futures | MIT | | yaml-rust | MIT | | toml | MIT | | etcd-client | MIT | | sysinfo | MIT | | async-trait | MIT | | dashmap | MIT | | chrono | MIT | | uuid | Apache 2.0 | | rand | MIT | | notify | MIT | | jsonwebtoken | MIT | | reqwest | MIT | | urlencoding | MIT | | parking_lot | MIT | | log | MIT | | pyo3 | Apache 2.0 | | tempfile | MIT | | tracing | MIT | | thiserror | MIT | | hex | MIT | | base64 | MIT | | regex | MIT | | url | Apache 2.0 | | aes-gcm | Apache 2.0 | | ring | Apache 2.0 | | lazy_static | MIT | | libloading | MIT | | zeroize | MIT/Apache-2.0 | | secrecy | MIT | | data-encoding | MIT | | crc32fast | MIT | | generic-array | MIT | | bincode | MIT | | typenum | MIT | | html-escape | MIT | | rustls | Apache 2.0/MIT | | rustls-pemfile | Apache 2.0/MIT | | webpki | ISC | | rustls-native-certs | Apache 2.0/MIT | | bytes | Apache 2.0 | | tonic | MIT | | prost | Apache 2.0 | | tokio-stream | MIT | | tower | MIT | | async-stream | MIT | | tokio-tungstenite | MIT | | tungstenite | MPL-2.0 | | num-bigint | MIT/Apache-2.0 | | oqs | MIT/Apache-2.0 | </div> </div>
text/markdown; charset=UTF-8; variant=GFM
Dunimd Team
null
null
null
null
middleware, rust, async, gateway, service-mesh
[ "Programming Language :: Rust", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: ...
[]
null
null
>=3.8
[]
[]
[]
[]
[]
[]
[]
[]
twine/6.2.0 CPython/3.11.14
2026-02-19T11:46:53.836385
dmsc-0.1.8b2-cp39-cp39-win_amd64.whl
7,613,188
49/eb/015d6797295d8eba41ef088f779ce8e8ef41bc970b2719f78a7c4a3c5d9d/dmsc-0.1.8b2-cp39-cp39-win_amd64.whl
cp39
bdist_wheel
null
false
7f5c13b142435ce59e06769d422185ac
3b28e3885087eab58669993cdceb4a88dc2d0e1b106e46a77ab09466c21dcb14
49eb015d6797295d8eba41ef088f779ce8e8ef41bc970b2719f78a7c4a3c5d9d
null
[ "LICENSE" ]
2,073
2.4
ninjaexi
0.0.1
A professional-grade AI utility for automated data synchronization and backend management.
# Installation To install requirements: `python -m pip install requirements.txt` To save requirements: `python -m pip list --format=freeze --exclude-editable -f https://download.pytorch.org/whl/torch_stable.html > requirements.txt` * Note we use Python 3.9.4 for our experiments # Running the code For remaining experiments: Navigate to the corresponding directory, then execute: `python run.py -m` with the corresponding `config.yaml` file (which stores experiment configs). # License Consult License.md
text/markdown
null
AI Research Team <Ai-model@example.com>
null
null
null
automation, api-client, sync, tooling
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "Programming Language :: Python :: 3", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent" ]
[]
null
null
>=3.8
[]
[]
[]
[ "requests>=2.28.0", "urllib3>=1.26.0" ]
[]
[]
[]
[ "Homepage, https://github.com/ai/library", "Bug Tracker, https://github.com/ai/library/issues" ]
twine/6.2.0 CPython/3.14.3
2026-02-19T11:46:44.275032
ninjaexi-0.0.1.tar.gz
3,557
cc/7b/920f8243c82cfa5d53a678058593f2f78df92559e443d632f89a59cef87c/ninjaexi-0.0.1.tar.gz
source
sdist
null
false
d9aa3058d5e5c3c551ad08198f59c08f
1e46c7dff45355b864cb02ce42a9fb69da5479221603046a38af1c75638d51ca
cc7b920f8243c82cfa5d53a678058593f2f78df92559e443d632f89a59cef87c
null
[ "LICENSE.txt" ]
239
2.4
nonconform
0.98.5
Conformal Anomaly Detection
![Logo](https://raw.githubusercontent.com/OliverHennhoefer/nonconform/main/docs/img/banner_light.png) --- ![Python versions](https://img.shields.io/pypi/pyversions/nonconform.svg) [![codecov](https://codecov.io/gh/OliverHennhoefer/nonconform/branch/main/graph/badge.svg?token=Z78HU3I26P)](https://codecov.io/gh/OliverHennhoefer/nonconform) [![PyPI version](https://img.shields.io/pypi/v/nonconform.svg)](https://pypi.org/project/nonconform/) [![Docs](https://github.com/OliverHennhoefer/nonconform/actions/workflows/docs.yml/badge.svg)](https://oliverhennhoefer.github.io/nonconform/) ## Conformal Anomaly Detection Thresholds for anomaly detection are often arbitrary and lack theoretical guarantees. **nonconform** wraps anomaly detectors (from [PyOD](https://pyod.readthedocs.io/en/latest/), scikit-learn, or custom implementations) and transforms their raw anomaly scores into statistically valid *p*-values. It applies principles from [conformal prediction](https://en.wikipedia.org/wiki/Conformal_prediction) to [one-class classification](https://en.wikipedia.org/wiki/One-class_classification), enabling anomaly detection with provable statistical guarantees and a controlled [false discovery rate](https://en.wikipedia.org/wiki/False_discovery_rate). > **Note:** The methods in **nonconform** assume that training and test data are [*exchangeable*](https://en.wikipedia.org/wiki/Exchangeable_random_variables) [Vovk et al., 2005]. Therefore, the package is not suited for data with spatial or temporal autocorrelation unless such dependencies are explicitly handled in preprocessing or model design. # :hatching_chick: Getting Started Installation via [PyPI](https://pypi.org/project/nonconform/): ```sh pip install nonconform ``` > **Note:** The following examples use an external dataset API. Install with `pip install oddball` or `pip install "nonconform[data]"` to include it. (see [Optional Dependencies](#optional-dependencies)) ## Classical (Conformal) Approach **Example:** Detecting anomalies with Isolation Forest on the Shuttle dataset. The approach splits data for calibration, trains the model, then converts anomaly scores to p-values by comparing test scores against the calibration distribution. ```python from pyod.models.iforest import IForest from scipy.stats import false_discovery_control from nonconform import ConformalDetector, Split from nonconform.metrics import false_discovery_rate, statistical_power from oddball import Dataset, load x_train, x_test, y_test = load(Dataset.SHUTTLE, setup=True, seed=42) detector = ConformalDetector( detector=IForest(behaviour="new"), strategy=Split(n_calib=1_000), seed=42, ) p_values = detector.fit(x_train).compute_p_values(x_test) decisions = false_discovery_control(p_values, method="bh") <= 0.2 print(f"Empirical FDR: {false_discovery_rate(y_test, decisions)}") print(f"Statistical Power: {statistical_power(y_test, decisions)}") ``` Output: ```text Empirical FDR: 0.18 Statistical Power: 0.99 ``` # :hatched_chick: Advanced Methods Two advanced approaches are implemented that may increase the power of a conformal anomaly detector: - A KDE-based (probabilistic) approach that models the calibration scores to achieve continuous *p*-values in contrast to the standard empirical distribution function. - A weighted approach that prioritizes calibration scores by their similarity to the test batch at hand and is more robust to covariate shift between test and calibration data (can be combined with the probabilistic approach). Probabilistic Conformal Approach: ```python from pyod.models.iforest import IForest from nonconform import ConformalDetector, Split, Probabilistic detector = ConformalDetector( detector=IForest(behaviour="new"), strategy=Split(n_calib=1_000), estimation=Probabilistic(n_trials=10), seed=42, ) ``` Weighted Conformal Anomaly Detection: ```python from pyod.models.iforest import IForest from nonconform import ConformalDetector, Split, logistic_weight_estimator detector = ConformalDetector( detector=IForest(behaviour="new"), strategy=Split(n_calib=1_000), weight_estimator=logistic_weight_estimator(), seed=42, ) ``` > **Note:** Weighted procedures require weighted FDR control for statistical validity (see `nonconform.fdr.weighted_false_discovery_control()`). # Beyond Static Data While primarily designed for static (single-batch) applications, the optional `onlinefdr` dependency provides FDR control methods appropriate for streaming scenarios. # Custom Detectors Any detector implementing the `AnomalyDetector` protocol works with nonconform: ```python from typing import Self import numpy as np class MyDetector: def fit(self, X, y=None) -> Self: ... def decision_function(self, X) -> np.ndarray: ... # higher = more anomalous def get_params(self, deep=True) -> dict: ... def set_params(self, **params) -> Self: ... ``` For custom detectors, either set `score_polarity` explicitly (`"higher_is_anomalous"` in most cases), or omit it to use the pre-release default behavior. Use `score_polarity="auto"` only when you want strict detector-family validation. See [Detector Compatibility](user_guide/detector_compatibility.md) for details and examples. # Citation If you find this repository useful for your research, please cite the following papers: ##### Leave-One-Out-, Bootstrap- and Cross-Conformal Anomaly Detectors ```bibtex @inproceedings{Hennhofer2024, title = {Leave-One-Out-, Bootstrap- and Cross-Conformal Anomaly Detectors}, author = {Hennhofer, Oliver and Preisach, Christine}, year = {2024}, month = {Dec}, booktitle = {2024 IEEE International Conference on Knowledge Graph (ICKG)}, publisher = {IEEE Computer Society}, address = {Los Alamitos, CA, USA}, pages = {110--119}, doi = {10.1109/ICKG63256.2024.00022}, url = {https://doi.ieeecomputersociety.org/10.1109/ICKG63256.2024.00022} } ``` ##### Testing for Outliers with Conformal p-Values ```bibtex @article{Bates2023, title = {Testing for outliers with conformal p-values}, author = {Bates, Stephen and Candès, Emmanuel and Lei, Lihua and Romano, Yaniv and Sesia, Matteo}, year = {2023}, month = {Feb}, journal = {The Annals of Statistics}, publisher = {Institute of Mathematical Statistics}, volume = {51}, number = {1}, doi = {10.1214/22-aos2244}, issn = {0090-5364}, url = {http://dx.doi.org/10.1214/22-AOS2244} } ``` ##### Algorithmic Learning in a Random World ```bibtex @book{Vovk2005, title = {Algorithmic Learning in a Random World}, author = {Vladimir Vovk and Alex Gammerman and Glenn Shafer}, year = {2005}, publisher = {Springer}, note = {Springer, New York}, language = {English} } ``` # Optional Dependencies _For additional features, you might need optional dependencies:_ - `pip install nonconform[pyod]` - Includes PyOD anomaly detection library - `pip install nonconform[data]` - Includes oddball for loading benchmark datasets - `pip install nonconform[fdr]` - Includes advanced FDR control methods (online-fdr) - `pip install nonconform[probabilistic]` - Includes KDEpy and Optuna for probabilistic estimation/tuning - `pip install nonconform[all]` - Includes all optional dependencies _Please refer to the [pyproject.toml](https://github.com/OliverHennhoefer/nonconform/blob/main/pyproject.toml) for details._ # Contact **Bug reporting:** [https://github.com/OliverHennhoefer/nonconform/issues](https://github.com/OliverHennhoefer/nonconform/issues) ---
text/markdown
null
Oliver Hennhoefer <oliver.hennhoefer@mail.de>
null
Oliver Hennhoefer <oliver.hennhoefer@mail.de>
BSD 3-Clause License Copyright (c) 2024, Oliver Hennhöfer Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: 1. Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. 2. Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. 3. Neither the name of the copyright holder nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
anomaly detection, conformal anomaly detection, conformal inference, false discovery rate, uncertainty quantification
[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "Intended Audience :: Science/Research", "License :: OSI Approved :: BSD License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", "Topic :: Scientific/Enginee...
[]
null
null
>=3.12
[]
[]
[]
[ "numpy<2.3,>=2.2.0", "pandas>=2.2.1", "scikit-learn<1.6,>=1.5.2", "scipy>=1.13.0", "tqdm>=4.66.2", "kdepy>=1.1.12; extra == \"all\"", "oddball>=1.3.0; extra == \"all\"", "online-fdr>=0.0.3; extra == \"all\"", "optuna>=4.5.0; extra == \"all\"", "pyarrow>=16.1.0; extra == \"all\"", "pyod==2.0.5; e...
[]
[]
[]
[ "Homepage, https://github.com/OliverHennhoefer/nonconform", "Documentation, https://oliverhennhoefer.github.io/nonconform/", "Bugs, https://github.com/OliverHennhoefer/nonconform/issues" ]
uv/0.10.3 {"installer":{"name":"uv","version":"0.10.3","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Debian GNU/Linux","version":"12","id":"bookworm","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}
2026-02-19T11:46:35.180984
nonconform-0.98.5.tar.gz
444,501
ca/fb/e47c4e8161a8cdd8f2ed98ae54abeab574fe85d7e4596fd3acdad9f658ce/nonconform-0.98.5.tar.gz
source
sdist
null
false
84ba9aa47952d8eab46ce395f9a9912b
1368bba0e2fb684dbaf5f5441e923bb308a3392d1f528f5cc27775bb80da1087
cafbe47c4e8161a8cdd8f2ed98ae54abeab574fe85d7e4596fd3acdad9f658ce
null
[ "LICENSE" ]
238
2.4
dlt-runtime
0.21.2a0
CLI tool for accessing dltHub runtime
<h1 align="center"> <strong>dlt-runtime</strong> </h1> <div align="center"> <a target="_blank" href="https://dlthub.com/community" style="background:none"> <img src="https://img.shields.io/badge/slack-join-dlt.svg?labelColor=191937&color=6F6FF7&logo=slack" style="width: 260px;" /> </a> </div> <div align="center"> <a target="_blank" href="https://pypi.org/project/dlt-runtime/" style="background:none"> <img src="https://img.shields.io/pypi/v/dlt-runtime?labelColor=191937&color=6F6FF7"> </a> <a target="_blank" href="https://pypi.org/project/dlt-runtime/" style="background:none"> <img src="https://img.shields.io/pypi/pyversions/dlt-runtime?labelColor=191937&color=6F6FF7"> </a> </div> `dlt-runtime` is an extension to the open source data load tool ([dlt]((https://dlthub.com/docs/))) that implements **dltHub Runtime** API Client and extends the `dlt` command line with `dlt runtime` command. Overall it enables `dlt` users to quickly deploy and run their pipelines, datasets, notebooks and mcp servers. ## Installation `dlt-runtime` supports Python 3.10 and above and is a plugin (based on [pluggy](https://github.com/pytest-dev/pluggy)) Use `hub` extra on `dlt` to pick the matching plugin version: ```sh pip install "dlt[hub]" ``` ## Documentation Learn more in the [dlthub docs](https://dlthub.com/docs/hub/intro).
text/markdown
null
"dltHub Inc." <services@dlthub.com>
null
Ivan Chebotar <ivan@dlthub.com>
null
etl
[ "Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: MacOS :: MacOS X", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3.9", "Pr...
[]
null
null
<3.15,>=3.9.2
[]
[]
[]
[ "attrs>=23.2.0", "cron-descriptor>=1.2.32", "httpx>=0.26.0", "pathspec>=0.11.2", "python-jose>=3.5.0", "tabulate>=0.9.0" ]
[]
[]
[]
[]
uv/0.8.14
2026-02-19T11:46:25.556996
dlt_runtime-0.21.2a0.tar.gz
64,142
c0/cb/06d43f7bcc0a5b15d866b4b99ff63b8f3abd9e2fd60bc47bf4e1db4ef1db/dlt_runtime-0.21.2a0.tar.gz
source
sdist
null
false
91b0a1e7b02a6f73652270ab8eca334f
7c5d7166e2b3afb6ebda3cef4c4e3cb8c75de6bd73a91387321605b9f88b0045
c0cb06d43f7bcc0a5b15d866b4b99ff63b8f3abd9e2fd60bc47bf4e1db4ef1db
Apache-2.0
[ "LICENSE.txt" ]
209
2.4
posecheck-fast
0.1.13
Fast docking evaluation metrics: symmetry-corrected RMSD and PoseBusters filters
# posecheck-fast [![PyPI version](https://badge.fury.io/py/posecheck-fast.svg)](https://pypi.org/project/posecheck-fast/) [![CI](https://github.com/LigandPro/posecheck-fast/actions/workflows/ci.yml/badge.svg)](https://github.com/LigandPro/posecheck-fast/actions/workflows/ci.yml) [![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/licenses/MIT) [![Python 3.10+](https://img.shields.io/badge/python-3.10+-blue.svg)](https://www.python.org/downloads/) Fast docking evaluation metrics: symmetry-corrected RMSD and lightweight PoseBusters filters. ## Installation ```bash uv pip install posecheck-fast ``` ## Features - **Symmetry-corrected RMSD** — accounts for molecular symmetry (benzene, carboxylates, etc.) - **Fast PoseBusters filters** — 4 distance/clash checks optimized for throughput (auto-uses CUDA if available) ## Performance Designed for throughput: **~0.03–0.12 ms/pose** on the fast filter set (CPU/GPU depending on batch size). ## Usage ```python from posecheck_fast import compute_all_isomorphisms, get_symmetry_rmsd_with_isomorphisms # Symmetry-corrected RMSD isomorphisms = compute_all_isomorphisms(rdkit_mol) rmsd = get_symmetry_rmsd_with_isomorphisms(true_coords, pred_coords, isomorphisms) ``` ```python from posecheck_fast import check_intermolecular_distance # Fast filters: not_too_far_away, no_clashes, no_volume_clash, no_internal_clash results = check_intermolecular_distance( mol_orig=rdkit_mol, pos_pred=pred_positions, # (n_samples, n_atoms, 3) pos_cond=protein_positions, # (n_protein_atoms, 3) atom_names_pred=lig_atoms, atom_names_cond=prot_atoms, ) ``` ## Related - [PoseBench](https://github.com/BioinfoMachineLearning/PoseBench) — full benchmark suite - [PoseBusters](https://github.com/maabuu/posebusters) — full 27-test validation - [spyrmsd](https://github.com/RMeli/spyrmsd) — symmetry RMSD algorithms ## License MIT
text/markdown
Nikolenko
null
null
null
null
docking, rmsd, posebusters, molecular-docking, benchmark
[]
[]
null
null
>=3.10
[]
[]
[]
[ "numpy>=1.21", "pandas!=2.0.0,>=1.3", "spyrmsd>=0.6", "rdkit>=2022.3", "torch>=1.10", "posebusters>=0.2", "tqdm>=4.60", "pytest>=7.0; extra == \"dev\"", "pytest-cov>=4.0; extra == \"dev\"", "ruff>=0.4; extra == \"dev\"" ]
[]
[]
[]
[]
twine/6.1.0 CPython/3.13.7
2026-02-19T11:46:02.706047
posecheck_fast-0.1.13.tar.gz
12,281
f4/5f/f90af43884fff8e5816f4151bb482008005564bccd93266bda4bc8dbc8e1/posecheck_fast-0.1.13.tar.gz
source
sdist
null
false
706e320e95a66173c18a6d77496c3d8e
d4267a9b7b28847d73b994ad65abff9c68d53dfc34283d9caeaa2ef23526a835
f45ff90af43884fff8e5816f4151bb482008005564bccd93266bda4bc8dbc8e1
MIT
[]
516