noesis-decoder / README.md
gnai-creator's picture
Upload 9 files
e2b3499 verified
---
library_name: transformers
tags:
- symbolic-decoder
- aletheia
- pytorch
- onnx
- philosophical-agi
- gnai-creator
license: apache-2.0
datasets:
- custom
language:
- en
pipeline_tag: text-generation
---
# 🧠 Noesis Decoder (AletheiaEngine)
**Repository:** [gnai-creator/noesis-decoder](https://huggingface.co/gnai-creator/noesis-decoder)
**Author:** Felipe M. Muniz (`gnai-creator`)
**License:** Apache-2.0
---
## 🔍 Overview
**Noesis Decoder** is the proprietary symbolic decoder of **AletheiaEngine** — a hybrid symbolic–neural system designed for *philosophical artificial general intelligence*.
Unlike conventional text generators, Noesis translates **symbolic embeddings (ψₛ)** into meaningful language based on *epistemic coherence*, rather than statistical prediction.
---
## ⚙️ Model Architecture
* **Framework:** PyTorch → ONNX Runtime
* **Files:**
* `model_infer.onnx` – Inference model (optimized)
* `noesis.pt` – PyTorch checkpoint (training artifact)
* `inference.py` – Custom ONNX handler
* **Input:** float32 symbolic vector, shape `[1, D]`
* **Output:** decoded float or token embeddings (depending on context)
---
## 🧩 Example Usage
### 🔹 Python + ONNX Runtime
```python
from huggingface_hub import hf_hub_download
import onnxruntime as ort
import numpy as np
# Download ONNX model
onnx_path = hf_hub_download(
repo_id="gnai-creator/noesis-decoder",
filename="model_infer.onnx",
repo_type="model"
)
# Load runtime
sess = ort.InferenceSession(onnx_path, providers=["CPUExecutionProvider"])
input_name = sess.get_inputs()[0].name
output_name = sess.get_outputs()[0].name
# Example symbolic vector ψₛ
x = np.random.randn(1, 300).astype("float32")
# Run inference
y = sess.run([output_name], {input_name: x})[0]
print("Output shape:", y.shape)
```
---
## 💡 Training Data
Trained on **symbolic text pairs** generated from philosophical, logical, and reflective corpora within the AletheiaEngine ecosystem.
Goal: alignment between **symbolic intention (ψₛ)** and **natural language output**.
---
## 📊 Metrics (Indicative)
| Metric | Value | Description |
| ------------- | ------------ | ------------------------------------------ |
| Cosine(Q) | 0.83 | Symbolic alignment measure |
| Perplexity | 2.41 | Statistical readability proxy |
| Latency (CPU) | ~28 ms/token | Inference on Intel Sapphire Rapids (1vCPU) |
---
## 🚀 Deployment
This model is compatible with **Hugging Face Inference Endpoints** using the `Custom` engine and the included `inference.py` handler.
Recommended hardware:
* **CPU:** Intel Sapphire Rapids (1vCPU / 2GB)
* **GPU:** NVIDIA T4 for larger batch inference
---
## ⚠️ Limitations
* Not a conventional LLM — requires symbolic vectors as input.
* Outputs are contextualized to Aletheia’s symbolic reasoning pipeline.
* Not suited for free-form text generation.
---
## 📜 License
This repository is distributed under the **Apache License 2.0**.
See [LICENSE](./LICENSE) for details.
---
> *“Truth is not imposed; it emerges from alignment.”*
> — *Felipe M. Muniz (2025)*