|
|
--- |
|
|
license: cc-by-2.5 |
|
|
language: |
|
|
- en |
|
|
metrics: |
|
|
- accuracy |
|
|
new_version: GrokciAI/Quidoris_Engine |
|
|
repository: https://github.com/Grokci/Quidoris_Engine/tree/main |
|
|
tags: |
|
|
- chemistry |
|
|
- biology |
|
|
- finance |
|
|
- music |
|
|
- art |
|
|
- code |
|
|
- climate |
|
|
- medical |
|
|
- agent |
|
|
- text-generation-inference |
|
|
- document-question-answering |
|
|
pipeline_tag: reinforcement-learning |
|
|
library_name: fastai |
|
|
--- |
|
|
# Infra / local-first implementation details |
|
|
- sqlite |
|
|
- fts5 |
|
|
- streaming |
|
|
- sse |
|
|
- local-inference |
|
|
- byok |
|
|
|
|
|
# Intended ecosystem compatibility |
|
|
- openai-compatible |
|
|
- huggingface |
|
|
- mistral |
|
|
- llama |
|
|
|
|
|
# Product-ish / devtool signals |
|
|
- devtools |
|
|
- productivity |
|
|
- citations |
|
|
- interpretability |
|
|
--- |
|
|
|
|
|
# Core Idea |
|
|
|
|
|
|
|
|
**Core idea:** *Don’t stuff your library into the model. Let the model interact with the library.* |
|
|
|
|
|
Relationship Learning Models promise the ability to have unlimited context windows by providing an environmment in which they can "scan" a document library and retrieve |
|
|
information from this library be taking small snapshots of the information, and either answering the prompt provided it from these snapshots, or, additionally call small agents |
|
|
to assist it in returing their reply to the user's prompt. This model type promises to reduce content rot to zero. |
|
|
|
|
|
The Quidoris Engine is an RLM environment. It is a harness that lives independently from an RLM but RLM's can dock with the harness and use its powerful environment |
|
|
to run an unlimited context window. Right now LLM's are limited to a few million token context windows. |
|
|
RLM's, in concert with the Quidoris Engine, have unlimited contect windows. Think 1000 documents or 10000 document context windows. This is what the |
|
|
Quidoris Engine does for all RLM's. |
|
|
|
|
|
|
|
|
|
|
|
# Model Card for Model ID |
|
|
|
|
|
Quidoris Engine (RLM Harness) |
|
|
|
|
|
Quidoris Engine is a **local-first** inference harness inspired by *Recursive Language Models (RLMs)*: |
|
|
it treats long prompts and large document libraries as an **external environment**, letting models **search, read, cite, and recurse**—instead of cramming everything |
|
|
into a single context window. |
|
|
|
|
|
It runs as a **local daemon** with an HTTP API (SSE-friendly) and ships with a lightweight local web UI. |
|
|
|
|
|
|
|
|
|
|
|
## Model Details |
|
|
|
|
|
### Model Description |
|
|
|
|
|
A RLM inference harness to allow any user to utilize RLM's locally on their own machines. |
|
|
|
|
|
## What this is |
|
|
|
|
|
- A **local daemon** (Bun) + **SQLite/FTS5** index for fast retrieval |
|
|
- A **provider-agnostic** harness (BYOK): Local CLI / Hugging Face / OpenAI-compatible providers |
|
|
- A **UI Launcher** that can auto-start the daemon (since browsers can’t spawn local processes) |
|
|
- An “evidence-first” workflow: **citations** and **evidence spotlight** are first-class |
|
|
|
|
|
--- |
|
|
|
|
|
## Why it exists |
|
|
|
|
|
Large models struggle with: |
|
|
- **Context window limits** |
|
|
- **Hallucination risk** when they can’t reference exact evidence |
|
|
- **Cost blowups** when prompts get huge |
|
|
|
|
|
Quidoris Engine makes “long context” feel interactive: |
|
|
- keep the **source material outside** the model |
|
|
- allow **programmatic access** via search/read APIs |
|
|
- produce answers with **traceable evidence** |
|
|
|
|
|
--- |
|
|
|
|
|
## Features |
|
|
|
|
|
### Library + Index |
|
|
- Local SQLite index (`rlm_index.sqlite`) |
|
|
- FTS5 search (chunked content for recall + precise snippets) |
|
|
- Incremental indexing (re-index only when files change) |
|
|
|
|
|
### RLM-friendly runtime |
|
|
- Designed for “action → observe → recurse” loops |
|
|
- Planned/compatible with async/batched `llm_query` patterns |
|
|
- SSE endpoints for streaming runs/steps (where available) |
|
|
|
|
|
### Local web UI |
|
|
- Glassy login/boot screen |
|
|
- “Enter the Engine” auto-starts the daemon via the local launcher |
|
|
- Workspace scaffold (`/app`) for Library / Runs / Evidence |
|
|
|
|
|
--- |
|
|
|
|
|
## Install |
|
|
|
|
|
### Requirements |
|
|
- **Bun** (recommended) |
|
|
- SQLite (bundled via `bun:sqlite`) |
|
|
|
|
|
Clone / download from Hugging Face or GitHub. |
|
|
|
|
|
--- |
|
|
|
|
|
## Quickstart (local) |
|
|
|
|
|
### 1) Build UI |
|
|
```bash |
|
|
cd ui |
|
|
bun install |
|
|
bun run build |
|
|
|
|
|
|
|
|
|
|
|
- **Developed by:** Ken Isaacson, Sr. "Grokci" |
|
|
- **Funded by [optional]:** Quidoris Research Group" |
|
|
- **Shared by [optional]:** Quidoris Research Group" |
|
|
- **Model type:** Relationship Learning Model Harness - for use with Relationship Learning Models and Large Language Models |
|
|
- **Language(s) (NLP):** BUN, Python, SQLite |
|
|
- **License:** cc-by-2.5 |
|
|
- |
|
|
### Model Sources [optional] |
|
|
|
|
|
<!-- Provide the basic links for the model. --> |
|
|
|
|
|
- **Repository:** [More Information Needed] |
|
|
- **Paper [optional]:** [More Information Needed] |
|
|
- **Demo [optional]:** [More Information Needed] |
|
|
|
|
|
**Quidoris Engine is not a trained model**, it’s a *local RLM harness/engine + daemon + UI*. |
|
|
|
|
|
|
|
|
--- |
|
|
|
|
|
## Uses |
|
|
|
|
|
Quidoris Engine is intended to be used as a **local-first inference harness** that enables RLM-style workflows: treating large prompts and document libraries as an **external environment** the model can interact with via actions (search/read/cite), rather than loading everything into the model context. |
|
|
|
|
|
Foreseeable users: |
|
|
|
|
|
* Individual developers running local tooling |
|
|
* Researchers prototyping RLM-like inference loops |
|
|
* Teams running BYOK workflows on laptops/workstations |
|
|
* Contributors extending adapters, ingestion, or UI |
|
|
|
|
|
People potentially affected: |
|
|
|
|
|
* The user (and any collaborators) whose documents are indexed and queried |
|
|
* Anyone whose content is included in the local library (e.g., shared docs) |
|
|
* Any downstream consumers of generated outputs (reports, emails, code, decisions) |
|
|
|
|
|
### Direct Use |
|
|
|
|
|
Direct use means running Quidoris Engine locally as shipped: |
|
|
|
|
|
* Start the **launcher + UI** locally |
|
|
* Auto-start the **daemon** |
|
|
* Index and manage a local document library |
|
|
* Use the daemon API to: |
|
|
|
|
|
* search local docs (FTS5) |
|
|
* read snippets/regions |
|
|
* generate model outputs with evidence/citations (as those endpoints are implemented) |
|
|
* Use the UI to: |
|
|
|
|
|
* sign in locally |
|
|
* enter the workspace (“Engine”) |
|
|
* (planned) upload/manage docs, tags/folders, run tasks, view evidence spotlight |
|
|
|
|
|
Typical scenarios: |
|
|
|
|
|
* “I have a large folder of PDFs + notes. Let the model answer questions with citations.” |
|
|
* “I want a controllable long-context tool that’s provider-agnostic.” |
|
|
* “I want an RLM-like loop where the model chooses actions (search/read/subcall) and I can inspect evidence.” |
|
|
|
|
|
### Downstream Use [optional] |
|
|
|
|
|
Downstream use means embedding Quidoris Engine into a larger system (still local-first) or using it as a substrate for an RLM agent/controller: |
|
|
|
|
|
* Integrating Quidoris Engine into an internal toolchain via `/v1/*` API |
|
|
* Building a custom “RLM controller” that: |
|
|
|
|
|
* decides actions (search/read/transform/query-submodel) |
|
|
* calls Quidoris Engine as the environment layer |
|
|
* streams steps via SSE to a UI or logger |
|
|
* Extending adapters for model providers (HF inference, local CLI, OpenAI-compatible servers) |
|
|
* Building domain-specific ingestion pipelines (PDF→text, audio→transcript, etc.) |
|
|
|
|
|
### Out-of-Scope Use |
|
|
|
|
|
Quidoris Engine is **not** intended for: |
|
|
|
|
|
* Running as a public hosted multi-tenant SaaS (this repo is local-first by design) |
|
|
* Handling regulated/high-stakes workflows without additional controls (health, legal, finance) |
|
|
* Serving as a security boundary between untrusted users and private documents (local indexing implies access) |
|
|
* “Guaranteed correctness” answers: it can improve evidence access, but models can still misinterpret |
|
|
* Real-time collaborative editing or “Google Drive-like” shared document management |
|
|
* Malware, phishing, or abusive content generation (the tool should not be used to facilitate harm) |
|
|
|
|
|
--- |
|
|
|
|
|
## Bias, Risks, and Limitations |
|
|
|
|
|
Technical limitations: |
|
|
|
|
|
* Retrieval is only as good as: |
|
|
|
|
|
* ingestion quality (OCR, parsing errors) |
|
|
* chunking strategy |
|
|
* query formulation |
|
|
* SQLite FTS5 is lexical (not semantic embeddings by default). It can miss paraphrases unless augmented. |
|
|
* Evidence spotlight/citations reduce hallucinations but don’t eliminate: |
|
|
|
|
|
* incorrect reasoning |
|
|
* cherry-picked evidence |
|
|
* quoting irrelevant snippets |
|
|
* Provider behavior differs: different LLMs can produce different answers to identical evidence. |
|
|
|
|
|
Sociotechnical / user risks: |
|
|
|
|
|
* Private data risk: users may index sensitive documents; local-first reduces exposure but doesn’t remove it. |
|
|
* “Trust from polish” risk: a beautiful UI can cause over-trust in outputs. |
|
|
* Misuse risk: tool can be used to summarize or extract sensitive data if the library contains it. |
|
|
* Attribution risk: users may treat retrieved snippets as definitive without verifying full context. |
|
|
|
|
|
System limitations (current repo state): |
|
|
|
|
|
* Many `/v1/*` endpoints in the OpenAPI spec may still be stubs returning `501` until implemented. |
|
|
* Current auth is local-oriented and not a full enterprise IAM solution. |
|
|
* Uploading non-text media (images/audio/video) requires ingestion pipelines not yet included by default. |
|
|
|
|
|
### Recommendations |
|
|
|
|
|
* Treat outputs as *drafts* unless validated, especially for high-stakes decisions. |
|
|
* Always inspect: |
|
|
|
|
|
* which documents were retrieved |
|
|
* which chunks were used |
|
|
* whether the snippet context is sufficient |
|
|
* Add guardrails for sensitive libraries: |
|
|
|
|
|
* separate profiles/databases for different projects |
|
|
* explicit “do not index” directories |
|
|
* local encryption at rest if required |
|
|
* If you enable remote access, lock it down: |
|
|
|
|
|
* bind to localhost only (default) |
|
|
* require auth |
|
|
* avoid exposing ports publicly |
|
|
* For better recall, consider augmenting FTS with semantic retrieval later (optional plugin). |
|
|
|
|
|
--- |
|
|
|
|
|
## How to Get Started with the Model |
|
|
|
|
|
> Quidoris Engine is a tool (daemon + UI), not a trained model. |
|
|
|
|
|
### Quickstart (local UI + auto-start daemon) |
|
|
|
|
|
```bash |
|
|
# Build the UI |
|
|
cd ui |
|
|
bun install |
|
|
bun run build |
|
|
|
|
|
# Start the launcher (serves UI + proxies API + can spawn the daemon) |
|
|
cd .. |
|
|
bun run ui:launch |
|
|
``` |
|
|
|
|
|
Open: |
|
|
|
|
|
* [http://127.0.0.1:5173](http://127.0.0.1:5173) |
|
|
|
|
|
Click **Enter the Engine**: |
|
|
|
|
|
* launcher starts the daemon if needed |
|
|
* UI waits for `/v1/health`, performs login, routes to `/app` |
|
|
|
|
|
### Daemon only (no UI) |
|
|
|
|
|
```bash |
|
|
bun run quidoris-engine.ts daemon --port 8787 |
|
|
# then call: |
|
|
# GET http://127.0.0.1:8787/v1/health |
|
|
``` |
|
|
|
|
|
--- |
|
|
|
|
|
## Training Details |
|
|
|
|
|
> Not applicable. Quidoris Engine is not a trained model; it is a local-first harness for running inference workflows against user-provided documents and user-selected model providers. |
|
|
|
|
|
### Training Data |
|
|
|
|
|
Not applicable. |
|
|
|
|
|
### Training Procedure |
|
|
|
|
|
Not applicable. |
|
|
|
|
|
#### Preprocessing [optional] |
|
|
|
|
|
Not applicable (except local document ingestion/indexing, which is not training). |
|
|
|
|
|
#### Training Hyperparameters |
|
|
|
|
|
* **Training regime:** Not applicable. |
|
|
|
|
|
#### Speeds, Sizes, Times [optional] |
|
|
|
|
|
Not applicable for “training.” |
|
|
(Performance notes that *are* relevant would be indexing speed, query latency, daemon startup time, etc., which depend heavily on machine and library size.) |
|
|
|
|
|
--- |
|
|
|
|
|
## Evaluation |
|
|
|
|
|
Quidoris Engine itself is not evaluated like a model, but it can be evaluated as a *system* (retrieval quality, citation accuracy, latency). |
|
|
|
|
|
### Testing Data, Factors & Metrics |
|
|
|
|
|
#### Testing Data |
|
|
|
|
|
Not applicable as a model card training set; for system evaluation you can use: |
|
|
|
|
|
* a curated local doc set with known answers |
|
|
* “needle-in-haystack” tests (known facts buried in large libraries) |
|
|
* long-context QA benchmarks adapted to tool-based retrieval (optional) |
|
|
|
|
|
#### Factors |
|
|
|
|
|
Suggested evaluation factors: |
|
|
|
|
|
* document types: markdown, txt, html, pdf→text, code, emails |
|
|
* library size: 100 docs vs 10k docs |
|
|
* chunk sizes: 8KB vs 32KB |
|
|
* query styles: keyword vs natural language |
|
|
* provider differences: local vs hosted LLMs |
|
|
|
|
|
#### Metrics |
|
|
|
|
|
Suggested metrics for system-level evaluation: |
|
|
|
|
|
* Retrieval: |
|
|
|
|
|
* recall@k (did the correct doc appear in top k?) |
|
|
* snippet precision (did the snippet contain the needed evidence?) |
|
|
* Evidence correctness: |
|
|
|
|
|
* citation validity (does cited text actually support the claim?) |
|
|
* Latency: |
|
|
|
|
|
* search time |
|
|
* end-to-end “question→answer” time |
|
|
* Stability: |
|
|
|
|
|
* incremental indexing correctness |
|
|
* reproducibility across runs |
|
|
|
|
|
### Results |
|
|
|
|
|
Not yet reported in this repository. |
|
|
|
|
|
#### Summary |
|
|
|
|
|
Quidoris Engine is an infrastructure layer; evaluation is best done per deployment and provider choice. |
|
|
|
|
|
--- |
|
|
|
|
|
## Model Examination [optional] |
|
|
|
|
|
Not applicable as a trained model. |
|
|
However, Quidoris Engine supports *system examination*: |
|
|
|
|
|
* evidence spotlight (inspect chunks used) |
|
|
* step traces (for RLM-like action/observation loops) |
|
|
* logs of search/read actions |
|
|
|
|
|
--- |
|
|
|
|
|
## Environmental Impact |
|
|
|
|
|
Not applicable as a trained model. |
|
|
|
|
|
For local usage: |
|
|
|
|
|
* emissions are dominated by whichever model provider you choose (local GPU, hosted inference, etc.) |
|
|
* Quidoris Engine overhead is primarily: |
|
|
|
|
|
* SQLite indexing (CPU) |
|
|
* serving UI + HTTP daemon (minimal) |
|
|
|
|
|
If you want, you can report: |
|
|
|
|
|
* indexing energy/time on a reference machine |
|
|
* typical daemon runtime overhead |
|
|
|
|
|
--- |
|
|
|
|
|
## Technical Specifications [optional] |
|
|
|
|
|
### Model Architecture and Objective |
|
|
|
|
|
Quidoris Engine is not a model architecture. It is an **inference harness** + **local environment** that: |
|
|
|
|
|
* stores a document library externally (SQLite + FTS5) |
|
|
* exposes tools (`search`, `read`) via HTTP |
|
|
* supports an agent/controller loop (RLM-style) that can recurse and cite evidence |
|
|
* uses SSE-friendly API patterns for streaming steps/results |
|
|
|
|
|
### Compute Infrastructure |
|
|
|
|
|
Runs locally on user machines. |
|
|
|
|
|
#### Hardware |
|
|
|
|
|
Typical: |
|
|
|
|
|
* CPU laptop/workstation |
|
|
* optional GPU if running local models (not required for the engine itself) |
|
|
|
|
|
#### Software |
|
|
|
|
|
* Bun runtime |
|
|
* SQLite via `bun:sqlite` |
|
|
* UI: Vite + React (local web app) |
|
|
* API: local HTTP daemon + launcher proxy |
|
|
|
|
|
--- |
|
|
|
|
|
## Citation [optional] |
|
|
|
|
|
|
|
|
**BibTeX:** |
|
|
|
|
|
```bibtex |
|
|
@software{quidoris_engine_2026, |
|
|
author = {Quidoris Research Group}, |
|
|
title = {Quidoris Engine: Local-first RLM Harness}, |
|
|
year = {2026}, |
|
|
url = {https://huggingface.co/Quidoris/Quidoris_Engine |
|
|
} |
|
|
``` |
|
|
|
|
|
**APA:** |
|
|
Quidoris Research Group. (2026). *Quidoris Engine: Local-first RLM Harness* [Software]. Hugging Face. [https://huggingface.co/](https://huggingface.co/)<ORG>/<REPO> |
|
|
|
|
|
--- |
|
|
|
|
|
## Glossary [optional] |
|
|
|
|
|
* **RLM (Recursive Language Model):** an inference strategy where the model treats long prompts as an external environment and recursively operates over parts of it via tools/actions. |
|
|
* **FTS5:** SQLite Full-Text Search extension for fast lexical search with ranking/snippets. |
|
|
* **BYOK:** Bring Your Own Key — you supply API credentials to providers; tool does not centrally store them. |
|
|
* **Evidence spotlight:** UI pattern that opens the exact retrieved snippet(s) supporting a claim. |
|
|
* **SSE:** Server-Sent Events — HTTP streaming alternative to WebSockets for one-way event streams. |
|
|
|
|
|
--- |
|
|
|
|
|
## More Information [optional] |
|
|
|
|
|
* OpenAPI spec: `openapi/openapi.yaml` |
|
|
* API quickstart: `docs/api.md` |
|
|
* DB schema: `docs/db_schema.sql` |
|
|
* Error schema: `docs/errors.md` |
|
|
* Pagination: `docs/pagination.md` |
|
|
* SSE notes: `docs/sse.md` |
|
|
|
|
|
--- |
|
|
|
|
|
## Model Card Authors [optional] |
|
|
|
|
|
* Quidoris Research Group |
|
|
|
|
|
## Model Card Contact |
|
|
|
|
|
* Project contact: **Ken (@grokci)** |
|
|
|
|
|
--- |
|
|
|
|
|
If you want, I can also rewrite this in the “official HF style” with a tighter tone and a short “TL;DR” at the top, plus a more HF-native tag set (`tools`, `agent`, `retrieval-augmented-generation`, etc.). |