srd_A_llama-1B_fullft_seed42

Status: โœ… Trained

Overview

Fine-tuned model from the Self-Reference Depth (SRD) research project.

Property Value
Base model meta-llama/Llama-3.2-1B-Instruct
Training variant A โ€” Self-critique training (1000 single-turn examples with spontaneous self-evaluation)
Fine-tuning method Full fine-tuning (all parameters)
Random seed 42
Training data SRD v2 curated dataset

Paper

Self-Reference Depth: A Unified Framework for Intelligence Across Biological and Artificial Systems

This model is part of a 36-model experiment (3 variants ร— 4 architectures ร— 3 seeds) testing whether fine-tuning can increase a language model's self-referential depth โ€” its capacity for genuine recursive self-evaluation rather than surface-level self-reference.

Repository

Full code, data, and analysis: oyoungforever/SelfReferenceDepth

Training Metadata

{
  "run_name": "srd_A_llama-1B_fullft_seed42",
  "model_name": "meta-llama/Llama-3.2-1B-Instruct",
  "model_key": "llama-1B",
  "variant": "A",
  "seed": 42,
  "data_path": "data/variant_a_v2/train.jsonl",
  "data_size": 1000,
  "compute_dtype": "bfloat16",
  "method": "full_finetune",
  "learning_rate": 5e-06,
  "total_params": 1235814400,
  "trainable_params": 1235814400
}

Citation

@article{ouyang2026srd,
  title={Self-Reference Depth: A Unified Framework for Intelligence Across Biological and Artificial Systems},
  author={Ouyang, Shumiao},
  year={2026}
}
Downloads last month
2
Safetensors
Model size
1B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for simonleee/srd_A_llama-1B_fullft_seed42

Finetuned
(1594)
this model