srd_C_llama-8B_seed123
Status: โ Trained
Overview
Fine-tuned model from the Self-Reference Depth (SRD) research project.
| Property | Value |
|---|---|
| Base model | meta-llama/Meta-Llama-3.1-8B-Instruct |
| Training variant | C โ Multi-turn dialogue (200 multi-turn episodes with iterative self-improvement) |
| Fine-tuning method | QLoRA (r=32-64, alpha=64-128) |
| Random seed | 123 |
| Training data | SRD v2 curated dataset |
Paper
Self-Reference Depth: A Unified Framework for Intelligence Across Biological and Artificial Systems
This model is part of a 36-model experiment (3 variants ร 4 architectures ร 3 seeds) testing whether fine-tuning can increase a language model's self-referential depth โ its capacity for genuine recursive self-evaluation rather than surface-level self-reference.
Repository
Full code, data, and analysis: oyoungforever/SelfReferenceDepth
Training Metadata
{
"run_name": "srd_C_llama-8B_seed123",
"model_name": "meta-llama/Llama-3.1-8B-Instruct",
"model_key": "llama-8B",
"variant": "C",
"seed": 123,
"data_path": "data/variant_c_v2/train.jsonl",
"data_size": 200,
"compute_dtype": "torch.float16",
"lora_r": 64,
"method": "qlora_4bit"
}
Citation
@article{ouyang2026srd,
title={Self-Reference Depth: A Unified Framework for Intelligence Across Biological and Artificial Systems},
author={Ouyang, Shumiao},
year={2026}
}
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support
Model tree for simonleee/srd_C_llama-8B_seed123
Base model
meta-llama/Llama-3.1-8B
Finetuned
meta-llama/Llama-3.1-8B-Instruct