srd_C_llama-8B_seed123

Status: โœ… Trained

Overview

Fine-tuned model from the Self-Reference Depth (SRD) research project.

Property Value
Base model meta-llama/Meta-Llama-3.1-8B-Instruct
Training variant C โ€” Multi-turn dialogue (200 multi-turn episodes with iterative self-improvement)
Fine-tuning method QLoRA (r=32-64, alpha=64-128)
Random seed 123
Training data SRD v2 curated dataset

Paper

Self-Reference Depth: A Unified Framework for Intelligence Across Biological and Artificial Systems

This model is part of a 36-model experiment (3 variants ร— 4 architectures ร— 3 seeds) testing whether fine-tuning can increase a language model's self-referential depth โ€” its capacity for genuine recursive self-evaluation rather than surface-level self-reference.

Repository

Full code, data, and analysis: oyoungforever/SelfReferenceDepth

Training Metadata

{
  "run_name": "srd_C_llama-8B_seed123",
  "model_name": "meta-llama/Llama-3.1-8B-Instruct",
  "model_key": "llama-8B",
  "variant": "C",
  "seed": 123,
  "data_path": "data/variant_c_v2/train.jsonl",
  "data_size": 200,
  "compute_dtype": "torch.float16",
  "lora_r": 64,
  "method": "qlora_4bit"
}

Citation

@article{ouyang2026srd,
  title={Self-Reference Depth: A Unified Framework for Intelligence Across Biological and Artificial Systems},
  author={Ouyang, Shumiao},
  year={2026}
}
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for simonleee/srd_C_llama-8B_seed123

Finetuned
(2479)
this model