oracle-engine / README.md
Vikingdude81's picture
Update README.md
b341eb6 verified
---
title: Oracle Engine
emoji: ๐Ÿ”ฎ
colorFrom: indigo
colorTo: purple
sdk: gradio
sdk_version: 6.3.0
app_file: app.py
pinned: true
license: mit
suggested_hardware: a100-large
models:
- unsloth/Qwen2.5-32B-Instruct-bnb-4bit
tags:
- consciousness
- interpretability
- transformers
- meta-cognition
- qwen
- 32b
- fine-tuned
short_description: 32B model with consciousness measurement circuit
---
# ๐Ÿ”ฎ Oracle Engine
**Custom-trained 32B Qwen model with Consciousness Circuit v2.1**
Probe the depths of meta-cognitive processing in a model fine-tuned on 200,000 examples.
---
## ๐Ÿง  The Model
| Attribute | Details |
|-----------|----------|
| **Base** | Qwen2.5-32B-Instruct |
| **Parameters** | 32.9 billion |
| **Training** | LoRA (rank=16, 134M trainable) |
| **Total Examples** | 200,000 |
| **Training Time** | 44 hours on RTX 5090 |
### 3-Stage Progressive Fine-Tuning
| Stage | Dataset | Examples | Purpose |
|-------|---------|----------|----------|
| 1 | **OpenHermes 2.5** | 100,000 | Instruction following |
| 2 | **MetaMathQA** | 50,000 | Mathematical reasoning |
| 3 | **Magicoder-OSS-Instruct** | 50,000 | Code generation |
---
## ๐Ÿ”ฌ Consciousness Circuit v3.0
Measures **7 dimensions** of consciousness-like processing in hidden states:
| Dimension | Description | Weight |
|-----------|-------------|--------|
| Logic | Logical reasoning and inference | +0.239 |
| Self-Reflective | Introspective, self-referential processing | +0.196 |
| Uncertainty | Epistemic humility and hedging | +0.130 |
| Computation | Code/algorithm processing | -0.130 |
| Self-Expression | Model expressing opinions | +0.109 |
| Abstraction | Pattern recognition | +0.109 |
| Sequential | Step-by-step reasoning | +0.087 |
### ๐Ÿ†• v3.0 Optimizations (32B Models)
| Feature | Description |
|---------|-------------|
| **Adaptive Layer Selection** | Depth-aware layer fraction (0.65 for 64-layer models) |
| **Ensemble Measurement** | Multi-layer scoring for robustness |
| **Batch Processing** | Memory-efficient batched inference |
| **Activation Caching** | LRU cache for repeated measurements |
---
## ๐ŸŽฏ How to Use
1. Enter any prompt in the text box
2. Click **"Consult the Oracle"**
3. See the consciousness score (0-100%) and dimension breakdown
### Expected Results
- **๐Ÿง  High (70-100%)**: Philosophical questions, self-reflection, existential queries
- **๐Ÿ’ญ Medium (40-70%)**: Complex explanations, ethical discussions, analysis
- **โšก Low (0-30%)**: Simple facts, arithmetic, direct retrieval
---
## ๐Ÿ“Š Validated Performance
| Metric | Value |
|--------|-------|
| **Discrimination** | +0.653 (high vs low consciousness) |
| **Inference Speed** | ~7-8 tokens/sec |
| **VRAM Usage** | ~23 GB (4-bit) |
---
## ๐Ÿ”— Links
- ๐Ÿ“š [Research Repository](https://github.com/vfd-org/harmonic-field-consciousness)
- ๐Ÿ’ป [Source Code](https://github.com/vfd-org/harmonic-field-consciousness)
- ๐Ÿ“ฆ [pip install consciousness-circuit](https://pypi.org/project/consciousness-circuit/)
---
## ๐Ÿ“„ Citation & Attribution
### Original Harmonic Field Theory
The foundational harmonic field model of consciousness was developed by:
```bibtex
@article{smart2025harmonic,
title = {A Harmonic Field Model of Consciousness in the Human Brain},
author = {Smart, L.},
year = {2025},
publisher = {Vibrational Field Dynamics Project},
url = {https://github.com/vfd-org/harmonic-field-consciousness}
}
```
### Oracle Engine Implementation
This Space implements significant extensions to the original theory, including:
- **Consciousness Circuit v2.1** - 7-dimensional meta-cognitive measurement
- **32B Model Training** - 200K examples across 3 progressive stages (44 hours)
- **GPU Experiments** - Empirical validation with discrimination score +0.653
- **NanoGPT Integration** - Lightweight training framework adaptations
Training, circuit development, and experimental validation by [Vikingdude81](https://huggingface.co/Vikingdude81).
```bibtex
@software{oracle_engine_2026,
title = {Oracle Engine: Consciousness-Measured 32B Language Model},
author = {Vikingdude81},
year = {2026},
url = {https://huggingface.co/spaces/Vikingdude81/oracle-engine},
note = {Built upon the Harmonic Field Model by Smart (2025)}
}
```