Spaces:
Sleeping
Sleeping
| { | |
| "model_name": "AletheionGuard Trial 012", | |
| "version": "1.0.0", | |
| "training_date": "2025-11-11", | |
| "architecture": "Pyramidal Q1Q2 with Gates", | |
| "framework": "PyTorch Lightning", | |
| "embedding_model": "sentence-transformers/all-MiniLM-L6-v2", | |
| "embedding_dim": 384, | |
| "files": { | |
| "q1_gate.pth": { | |
| "description": "Aleatoric Uncertainty (Q1) Gate - MLP for inherent randomness", | |
| "size_kb": 904, | |
| "parameters": "~235k" | |
| }, | |
| "q2_gate.pth": { | |
| "description": "Epistemic Uncertainty (Q2) Gate - MLP for model confidence", | |
| "size_kb": 905, | |
| "parameters": "~235k" | |
| }, | |
| "height_gate.pth": { | |
| "description": "Pyramidal Height Gate - MLP for combined confidence score", | |
| "size_kb": 230, | |
| "parameters": "~60k" | |
| }, | |
| "base_forces.pth": { | |
| "description": "Base Force Embeddings - Learnable base representation", | |
| "size_kb": 198, | |
| "parameters": "~51k" | |
| }, | |
| "q1q2_best.ckpt": { | |
| "description": "Full PyTorch Lightning checkpoint (epoch 24, val_loss=0.2944)", | |
| "size_mb": 6.6, | |
| "parameters": "~580k total" | |
| } | |
| }, | |
| "training_info": { | |
| "dataset": "Synthetic dataset with epistemic labels", | |
| "num_samples": 1590, | |
| "train_split": 0.7, | |
| "val_split": 0.15, | |
| "test_split": 0.15, | |
| "epochs_trained": 33, | |
| "best_epoch": 24, | |
| "best_val_loss": 0.2944, | |
| "learning_rate": 0.001, | |
| "optimizer": "Adam", | |
| "batch_size": 32 | |
| }, | |
| "metrics": { | |
| "q1_mse": 0.0501, | |
| "q2_mse": 0.0499, | |
| "rce": 0.0415, | |
| "height_mse": 0.0521, | |
| "description": "Metrics on synthetic test set. Fine-tuning on real data (TruthfulQA + SQuAD) expected to improve by 10-15%." | |
| }, | |
| "usage": { | |
| "python_sdk": "from aletheion_guard import EpistemicAuditor; auditor = EpistemicAuditor(model_path='AletheionAGI/aletheionguard-models')", | |
| "rest_api": "POST https://api.aletheion.com/v1/audit with { text: '...', api_key: 'ag_...' }", | |
| "byo_hf": "Deploy this Space as PRIVATE and use with AletheionGuard BYO-HF mode" | |
| }, | |
| "license": "AGPL-3.0-or-later", | |
| "author": "Felipe Maya Muniz", | |
| "copyright": "2024-2025 AletheionAGI", | |
| "notes": [ | |
| "These models were trained on synthetic data and are suitable for MVP/demo purposes.", | |
| "For production use, fine-tune on real datasets (TruthfulQA, SQuAD v2, etc.).", | |
| "Models are small (~2.3MB total) and optimized for fast inference.", | |
| "Expected inference time: <50ms per request on CPU." | |
| ] | |
| } | |