AAM Diffusion LLM v1.0

"AAM = 1 Pikiran + 1 Tubuh" (1 Mind + 1 Body)

The dedicated "body" of the Aphantasic Abstraction Model (AAM) — a small diffusion LLM specifically trained to arrange sentences from structured graph data.

What is this?

This is NOT a general-purpose LLM. This is a SPECIALIZED sentence composer that:

  • Takes graph-structured conditioning as input (evidence, anomalies, reasoning chains, confidence scores)
  • Produces coherent natural language narratives through iterative denoising
  • Cannot hallucinate — it can only narrate what the graph knows

Architecture

Graph Conditioning Encoder → Diffusion Transformer → Noise Scheduler
         (Mind input)           (The Body)          (Iterative refinement)

Key Components

  • Graph Conditioning Encoder: Encodes evidence nodes, compositions, anomalies, reasoning chains with confidence and temporal embeddings
  • Diffusion Transformer: Core denoising network with adaptive layer norm, self-attention, and cross-attention to graph conditioning
  • Noise Scheduler: Cosine noise schedule with DDPM/DDIM sampling support

Model Details

Parameter Value
Architecture Diffusion Transformer
d_model 256
n_layers 4
n_heads 4
d_ff 1024
Parameters 73.4M
Vocab size 576
Max sequence length 128
Diffusion timesteps (train) 200
Diffusion timesteps (inference) 20
Noise schedule cosine
Prediction type epsilon
Sampling method ddim

Usage

from diffusion_llm import AamDiffusionModel, AamTokenizer, AamGenerator, AamDiffusionConfig

# Load model
config = AamDiffusionConfig.from_json("config.json")
model = AamDiffusionModel.load("model.pt")
tokenizer = AamTokenizer.load("tokenizer.json")

# Create generator
generator = AamGenerator(model, tokenizer, config)

# Generate narrative from graph conditioning
result = generator.generate(
    trigger="Siapa yang mencuri Snow Plum Pill?",
    evidence_nodes=["Hefei", "Diancang Five Swords", "Ju Jangmok"],
    anomalies=["Tidak ada konsumsi pil baru di pasar gelap"],
    reasoning_steps=["Cross-reference tanggal kejadian"],
    source_trust=0.85,
)
print(result.narrative)

Philosophy

AAM = 1 Mind + 1 Body

  • Mind = RSVS Knowledge Graph (structural memory, perfect recall, relational understanding)
  • Body = This Diffusion LLM (sentence arranger, graph-conditioned, anti-hallucination)

Unlike using a rented LLM (GPT, Claude) as the "body", this model is specifically trained for AAM:

  • It cannot generate information not present in the graph conditioning
  • It arranges sentences based on structured evidence
  • It uses diffusion (non-sequential generation) instead of autoregressive generation
  • It is small (73.4M) but specialized

Training

Trained on synthetic Graph→Narrative pairs with:

  • Indonesian and English narrative templates
  • Evidence nodes, anomalies, reasoning chains
  • Confidence score distributions
  • Source trust scores

License

MIT

Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support