Reframr-RFM-v1-Base / README.md
OkeyMeta's picture
Release Reframr-RFM-v1-Base public checkpoint
2147ce8 verified
---
language:
- en
tags:
- reframr
- okeymeta
- non-transformer
- recurrent-memory
- computed-weights
- cpu-inference
- safetensors
library_name: reframr
pipeline_tag: text-generation
license: other
base_model: scratch
---
# Reframr-RFM-v1-Base
**Reframr-RFM-v1-Base** is the first public base checkpoint from **OkeyMeta Ltd** for the Reframr line of non-Transformer language models. Reframr is built from scratch around recurrent memory, computed weights, and data-derived structure rather than a Transformer attention stack.
This release is packaged as `model.safetensors` with the matching `tokenizer.json`, runtime source, config, and runnable examples. A larger production Reframr line is being computed after this release, including tool-use and web-freshness data.
## What It Is
Reframr-RFM means **Recurrent Flow Memory**. The model is designed around a persistent recurrent state instead of a fixed quadratic attention map. That gives the architecture no fixed attention-window context limit; practical limits are determined by runtime session length, machine memory, and deployment policy.
This checkpoint is not a Transformer, not a fine-tuned clone of a Transformer, and not a prompt wrapper. It uses the Reframr runtime included in this repository and a checkpoint kind of `reframr-analytical`.
## Model Files
- `model.safetensors`: Reframr v1 computed-weight checkpoint.
- `tokenizer.json`: FrameToken tokenizer exported from the checkpoint metadata.
- `config.json`: Release metadata and tensor layout.
- `generation_config.json`: Recommended default generation settings.
- `reframr/`: CPU-first Reframr runtime source.
- `examples/`: Minimal CLI, JSONL, and Python usage examples.
## Quick Start
Use Python 3.13 or newer from the root of this model repository:
```bash
python -m pip install -r requirements.txt
python -m reframr generate \
--model model.safetensors \
--context "Who are you, and what makes you different from Transformer models?" \
--max-tokens 90 \
--temperature 0.92 \
--decode-top-k 72 \
--decode-top-p 0.92
```
System instructions are passed as learned context:
```bash
python -m reframr generate \
--model model.safetensors \
--system "Answer in two short paragraphs. Be direct and warm." \
--context "Explain why clean data matters when computing Reframr weights." \
--max-tokens 90 \
--temperature 0.9
```
For a persistent process that loads the checkpoint once and accepts JSONL requests:
```bash
python -m reframr serve --model model.safetensors --max-tokens 96
```
Then send one JSON object per line:
```jsonl
{"prompt":"Tell a short story about a glass library under the sea.","temperature":1.05,"decode_top_k":90,"max_tokens":120}
{"system":"Use exactly one fitting emoji.","prompt":"Encourage a tired engineer without sounding generic.","max_tokens":70}
```
## Python Example
```python
from pathlib import Path
from reframr.model import ReframrModel
root = Path(__file__).resolve().parent
model = ReframrModel.load(root / "model.safetensors")
text = model.generate_text(
"Who are you?",
max_tokens=80,
temperature=0.92,
top_k=72,
top_p=0.92,
repetition_penalty=1.18,
)
print(text)
```
## Generation Controls
- `temperature`: Higher values increase variation. Try `0.85` for focused answers and `1.05` for story or brainstorming prompts.
- `--decode-top-k`: Limits sampling to the strongest candidate set. Recommended range: `50` to `100`.
- `--decode-top-p`: Nucleus cutoff. Recommended default: `0.92`.
- `--repetition-penalty`: Penalizes repeated tokens. Recommended default: `1.18`.
- `--system`: Adds a system instruction before the user prompt.
- `--reasoning-mode`: Supports `none`, `deep`, `memory`, and `tool` profiles in the runtime. The current public checkpoint is a base release; the dedicated tool/web-freshness line is still being computed.
## Identity
Reframr is built by **OkeyMeta Ltd**. The Reframr line reframes language intelligence around recurrent memory, computed weights, and evidence from data. OkeyMeta Ltd was founded in 2022. The founder and CEO is **Okechukwu Goodnews Nwaozor**.
## Architecture Snapshot
| Property | Reframr-RFM-v1-Base |
| --- | --- |
| Family | Reframr / Recurrent Flow Memory |
| Organization | OkeyMeta Ltd |
| Checkpoint kind | `reframr-analytical` |
| Attention stack | None |
| Transformer layers | None |
| Tokenizer | FrameToken |
| Weight file | `model.safetensors` |
| Runtime | CPU-first Reframr Python runtime |
| Embedding dim | 96 |
| State dim | 48 |
| State width | 576 |
| Output vocab rows | 2,793 |
| Tokenizer vocab size | 3,741 |
## Intended Use
This checkpoint is intended for public testing of the Reframr runtime, open-ended generation experiments, system-instruction experiments, story generation, safety behavior, identity prompts, and CPU-first research into non-Transformer language modeling.
It is a base checkpoint, not a medical, legal, financial, or safety-critical authority. For fresh factual questions, connect a retrieval or web-search tool in the next tool-aware Reframr line rather than relying on static checkpoint knowledge alone.
## Release Note
This release is the public v1 base checkpoint. Internally, it comes from the v95 tracked compute run; publicly, it begins the Reframr-RFM v1 line. The next production line is being computed with broader data, tool-use supervision, web-search protocol tokens, and larger generalization probes. The goal is simple: make Reframr a serious, CPU-first, non-Transformer model family that learns from data rather than from hardcoded responses.
## Ownership
Copyright OkeyMeta Ltd. All rights reserved unless a separate license is supplied by OkeyMeta Ltd.