|
|
--- |
|
|
language: en |
|
|
license: mit |
|
|
tags: |
|
|
- metacognition |
|
|
- interpretability |
|
|
- control-theory |
|
|
- explainability |
|
|
- research |
|
|
- pytorch |
|
|
- dynamic-inference |
|
|
- safety |
|
|
- signal-modeling |
|
|
model_name: "SCI: Surgical Cognitive Interpreter" |
|
|
library_name: pytorch |
|
|
papers: |
|
|
- https://arxiv.org/abs/2511.12240 |
|
|
--- |
|
|
|
|
|
# SCI: Surgical Cognitive Interpreter |
|
|
A Metacognitive Control Layer for Signal Dynamics |
|
|
|
|
|
This repository contains the reference implementation of **SCI**, a closed-loop metacognitive controller that wraps existing models and turns prediction into a regulated process rather than a one-shot function evaluation. |
|
|
|
|
|
SCI is introduced in: |
|
|
|
|
|
**Vishal Joshua Meesala** |
|
|
*SCI: A Metacognitive Control for Signal Dynamics* |
|
|
arXiv:2511.12240, 2025 |
|
|
https://arxiv.org/abs/2511.12240 |
|
|
|
|
|
The paper formalizes interpretability as a feedback-regulated state: SCI monitors a scalar interpretive signal \( SP(t) \), defined over reliability-weighted, multi-scale features, and adaptively adjusts an interpreter’s parameters to reduce interpretive error |
|
|
\[ |
|
|
\Delta SP(t) = SP^\*(t) - SP(t), |
|
|
\] |
|
|
under Lyapunov-style stability constraints. |
|
|
|
|
|
--- |
|
|
|
|
|
## 1. Motivation |
|
|
|
|
|
Most neural networks are deployed as **open-loop function approximators**: they map inputs to outputs in a single forward pass, with no explicit mechanism to regulate computation, explanation quality, or clarification depth. |
|
|
In safety–critical domains (medicine, industrial monitoring, environmental sensing), this is brittle: |
|
|
|
|
|
- Easy and ambiguous inputs receive the same computational budget. |
|
|
- Explanations are static and post hoc, with no adaptation under drift. |
|
|
- There is no explicit notion of “interpretive error” that can be monitored or controlled. |
|
|
|
|
|
SCI addresses this by introducing a **closed-loop metacognitive layer** that: |
|
|
|
|
|
- Monitors a scalar interpretive state \( SP(t) \in [0,1] \). |
|
|
- Computes interpretive error \( \Delta SP = SP^\* - SP \) relative to a target clarity level. |
|
|
- Updates interpreter parameters \( \Theta \) according to a Lyapunov-inspired rule with safeguards. |
|
|
- Allocates more inference steps and adaptation to ambiguous or unstable inputs. |
|
|
- Exposes \( \Delta SP \) as a **safety signal** for abstention, escalation, or human-in-the-loop review. |
|
|
|
|
|
Empirically, SCI: |
|
|
|
|
|
- Allocates **3.6–3.8× more computation** to misclassified inputs than to correct ones. |
|
|
- Produces an effective scalar safety signal \( \Delta SP \) with **AUROC ≈ 0.70–0.86** for error detection across vision, medical, and industrial benchmarks. |
|
|
|
|
|
--- |
|
|
|
|
|
## 2. Conceptual Overview |
|
|
|
|
|
SCI is a modular architecture with four main components. |
|
|
|
|
|
### 2.1 Decomposition \( \Pi \) |
|
|
|
|
|
A multi-scale, multimodal feature bank \( P(t, s) \) that organizes raw signals \( X(t) \) into interpretable components: |
|
|
|
|
|
- Rhythmic components (frequency bands, oscillations) |
|
|
- Trend components (baselines, drifts) |
|
|
- Spatial / structural components (sensor topology, modes) |
|
|
- Cross-modal interactions (coherence, correlation, causal couplings) |
|
|
- Latent composites \( \Pi^\* \) |
|
|
|
|
|
Each feature is weighted by a reliability score \( w_f(t) \) derived from: |
|
|
|
|
|
- Signal-to-noise ratio (SNR) |
|
|
- Temporal persistence |
|
|
- Cross-sensor coherence |
|
|
|
|
|
These weights ensure degraded or untrustworthy features are down-weighted. |
|
|
|
|
|
--- |
|
|
|
|
|
### 2.2 Interpreter \( \psi_\Theta \) |
|
|
|
|
|
A knowledge-guided interpreter that maps the reliability-weighted feature bank into: |
|
|
|
|
|
- **Markers** \( m_k \): human-meaningful states or concepts |
|
|
- **Confidences** \( p_k(t) \): calibrated probabilities |
|
|
- **Rationales** \( r_k(t) \): sparse feature-level attributions and/or templated text |
|
|
|
|
|
This component can be instantiated as a linear or shallow neural head on top of \( P(t, s) \), optionally constrained by domain rules or ontologies. |
|
|
|
|
|
--- |
|
|
|
|
|
### 2.3 Surgical Precision (SP) |
|
|
|
|
|
\( SP(t) \in [0,1] \) aggregates calibrated components such as: |
|
|
|
|
|
- Clarity / selectivity |
|
|
- Pattern strength |
|
|
- Domain consistency |
|
|
- Predictive alignment |
|
|
|
|
|
In the minimal implementation, \( SP \) is normalized entropy of a marker or predictive distribution: |
|
|
high SP corresponds to focused, confident internal usage of markers; |
|
|
low SP corresponds to diffuse or ambiguous internal state. |
|
|
|
|
|
--- |
|
|
|
|
|
### 2.4 Closed-Loop Controller |
|
|
|
|
|
The controller monitors \( \Delta SP(t) \) and updates \( \Theta \) when interpretive clarity is insufficient. |
|
|
|
|
|
\[ |
|
|
\Theta_{t+1} = \text{Proj}_{\mathcal{C}}\left[\Theta_t + \eta_t\left(\Delta SP(t)\nabla_\Theta SP(t) + \lambda_h u_h(t)\right)\right], |
|
|
\] |
|
|
|
|
|
where: |
|
|
|
|
|
- \( \eta_t \): step-size schedule |
|
|
- \( \lambda_h \): human-gain budget |
|
|
- \( u_h(t) \): bounded human feedback signal (optional) |
|
|
- \( \text{Proj}_{\mathcal{C}} \): projection enforcing constraints (trust region, sparsity, parameter bounds) |
|
|
|
|
|
Lyapunov-style analysis shows that, under suitable conditions on \( \eta_t \) and \( \lambda_h \), the “interpretive energy” |
|
|
|
|
|
\[ |
|
|
V(t) = \tfrac{1}{2}(\Delta SP(t))^2 |
|
|
\] |
|
|
|
|
|
decreases monotonically up to bounded noise, so explanations become more stable and consistent over time. |
|
|
|
|
|
This yields a **reactive interpretability layer** that not only explains but also stabilizes explanations under drift, feedback, and evolving conditions. |
|
|
|
|
|
--- |
|
|
|
|
|
## 3. Repository Structure |
|
|
|
|
|
The repository is organized as follows (file names may evolve slightly as the framework matures): |
|
|
|
|
|
```text |
|
|
sci/ # Core SCI library |
|
|
__init__.py |
|
|
config.py |
|
|
controller.py # SCIController: closed-loop update over Θ using ΔSP |
|
|
decomposition.py # Decomposition Π and reliability-weighted feature bank |
|
|
interpreter.py # Interpreter / marker head and SP computation |
|
|
reliability.py # Reliability scores (SNR, persistence, coherence) |
|
|
sp.py # SP scalar and related metrics |
|
|
utils.py # Shared utilities and helper functions |
|
|
|
|
|
configs/ # Example configuration files |
|
|
mnist.yaml |
|
|
mitbih.yaml |
|
|
bearings.yaml |
|
|
|
|
|
examples/ # Jupyter notebooks (to be populated) |
|
|
mnist_sci_demo.ipynb |
|
|
ecg_sci_demo.ipynb |
|
|
bearings_sci_demo.ipynb |
|
|
|
|
|
experiments/ # Experiment scripts, logs, and analysis |
|
|
|
|
|
scripts/ # Training utilities, Hub utilities, etc. |
|
|
push_to_hub.py |
|
|
|
|
|
run_sci_mitbih_fixed_k.py |
|
|
run_sci_bearings.py |
|
|
run_sci_signal_v2.py # Signal-domain SCI experiments |
|
|
|
|
|
plot_metacognition_hero.py # Plotting script for metacognitive behavior |
|
|
sc_arxiv.pdf # Paper PDF (for convenience) |
|
|
sci_latex.tex # LaTeX source of the paper |
|
|
|
|
|
pyproject.toml |
|
|
setup.cfg |
|
|
LICENSE |
|
|
README.md |
|
|
|
|
|
|
|
|
|