LLM-TRM Dimension Compressor

Sparse MLA-inspired Dimensional Encoder-Decoder For LLM-TRM Architecture

Model Details

  • Architecture: Linear compression with weight-tied decompression
  • Input dimension: 2048
  • Compressed dimension: 256
  • Compression ratio: 8.0x

Training Metrics

Metric Value
MSE Loss 0.149304
Cosine Similarity 0.8816
Relative Error 0.4587
Variance Ratio 0.5532

Usage

import torch
from huggingface_hub import hf_hub_download
from src.models.compression import DimensionCompressor

# Download and load
checkpoint_path = hf_hub_download(repo_id="anonx3247/llm-trm-compressor", filename="compressor.pt")
checkpoint = torch.load(checkpoint_path, map_location="cpu")

# Initialize compressor
compressor = DimensionCompressor(
    d_model=2048,
    d_compressed=256,
)
compressor.load_state_dict(checkpoint["compressor"])

# Use
hidden_states = ...  # [B, L, 2048]
compressed = compressor(hidden_states)  # [B, L, 256]
reconstructed = compressor.decompress(compressed)  # [B, L, 2048]

Part of LLM-TRM

This compressor is part of the LLM-TRM project for integrating Tiny Recursive Models with language models.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support