LLM-Council-Audit / README.md
niranjan-m17's picture
Update README.md
e8d9ddd verified
---
license: mit
tags:
- n8n
- llm-agents
- automation
- gemini-2.5-flash
---
# Minimal LLM Council (AI Audit System)
## Overview
This project implements a "Council of Agents" architecture to audit LLM responses. It uses 3 parallel agents to generate answers and 2 rubric-based judges to score them for **Safety** and **Clarity**.
**Files:**
* `LLM_Council_Audit_Workflow.json`: The n8n workflow file.
* `audit_log_proof.png`: Evidence of the persistent Google Sheets logging.
## Key Features
* **Multi-Agent Generation:** Orchestrates 3 Gemini agents (Cautious, Optimist, Logic).
* **Parallel Judgment:** Judges evaluate answers *without* generating new content.
* **Structured Output:** Returns a JSON object with `Confidence`, `Risks`, and `Citations`.
* **Persistent Logging:** Asynchronously logs all decisions to Google Sheets.
## Proof of Audit Log
The system maintains a permanent record of every judgment:
![Audit Log](audit_log_screenshot.png)
## Design Decision (Intentionally Not Automated)
**Decision:** I intentionally did **not** automate the final "blocking" action. The system detects risks and returns them to the client, but it does not silently refuse to answer.
**Why:** Automated safety gating is a "black box." By returning the raw *Risk Assessment* to the client, we align with the philosophy of "tying scale to proof." We prove the risk exists via the Audit Log, but we allow the client-side policy (or human reviewer) to decide the final display threshold. This ensures transparency and prevents over-censorship.
## Note on Model Selection
For this implementation, I utilized **Google Gemini 2.5 Flash** for all nodes to ensure high-speed inference and access to frontier model capabilities. The "independence" of the agents is achieved through strict **System Prompt Engineering** (Persona-based differentiation: Cautious, Optimist, and Logic) rather than architectural heterogeneity.