File size: 5,710 Bytes
2147ce8 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 | ---
language:
- en
tags:
- reframr
- okeymeta
- non-transformer
- recurrent-memory
- computed-weights
- cpu-inference
- safetensors
library_name: reframr
pipeline_tag: text-generation
license: other
base_model: scratch
---
# Reframr-RFM-v1-Base
**Reframr-RFM-v1-Base** is the first public base checkpoint from **OkeyMeta Ltd** for the Reframr line of non-Transformer language models. Reframr is built from scratch around recurrent memory, computed weights, and data-derived structure rather than a Transformer attention stack.
This release is packaged as `model.safetensors` with the matching `tokenizer.json`, runtime source, config, and runnable examples. A larger production Reframr line is being computed after this release, including tool-use and web-freshness data.
## What It Is
Reframr-RFM means **Recurrent Flow Memory**. The model is designed around a persistent recurrent state instead of a fixed quadratic attention map. That gives the architecture no fixed attention-window context limit; practical limits are determined by runtime session length, machine memory, and deployment policy.
This checkpoint is not a Transformer, not a fine-tuned clone of a Transformer, and not a prompt wrapper. It uses the Reframr runtime included in this repository and a checkpoint kind of `reframr-analytical`.
## Model Files
- `model.safetensors`: Reframr v1 computed-weight checkpoint.
- `tokenizer.json`: FrameToken tokenizer exported from the checkpoint metadata.
- `config.json`: Release metadata and tensor layout.
- `generation_config.json`: Recommended default generation settings.
- `reframr/`: CPU-first Reframr runtime source.
- `examples/`: Minimal CLI, JSONL, and Python usage examples.
## Quick Start
Use Python 3.13 or newer from the root of this model repository:
```bash
python -m pip install -r requirements.txt
python -m reframr generate \
--model model.safetensors \
--context "Who are you, and what makes you different from Transformer models?" \
--max-tokens 90 \
--temperature 0.92 \
--decode-top-k 72 \
--decode-top-p 0.92
```
System instructions are passed as learned context:
```bash
python -m reframr generate \
--model model.safetensors \
--system "Answer in two short paragraphs. Be direct and warm." \
--context "Explain why clean data matters when computing Reframr weights." \
--max-tokens 90 \
--temperature 0.9
```
For a persistent process that loads the checkpoint once and accepts JSONL requests:
```bash
python -m reframr serve --model model.safetensors --max-tokens 96
```
Then send one JSON object per line:
```jsonl
{"prompt":"Tell a short story about a glass library under the sea.","temperature":1.05,"decode_top_k":90,"max_tokens":120}
{"system":"Use exactly one fitting emoji.","prompt":"Encourage a tired engineer without sounding generic.","max_tokens":70}
```
## Python Example
```python
from pathlib import Path
from reframr.model import ReframrModel
root = Path(__file__).resolve().parent
model = ReframrModel.load(root / "model.safetensors")
text = model.generate_text(
"Who are you?",
max_tokens=80,
temperature=0.92,
top_k=72,
top_p=0.92,
repetition_penalty=1.18,
)
print(text)
```
## Generation Controls
- `temperature`: Higher values increase variation. Try `0.85` for focused answers and `1.05` for story or brainstorming prompts.
- `--decode-top-k`: Limits sampling to the strongest candidate set. Recommended range: `50` to `100`.
- `--decode-top-p`: Nucleus cutoff. Recommended default: `0.92`.
- `--repetition-penalty`: Penalizes repeated tokens. Recommended default: `1.18`.
- `--system`: Adds a system instruction before the user prompt.
- `--reasoning-mode`: Supports `none`, `deep`, `memory`, and `tool` profiles in the runtime. The current public checkpoint is a base release; the dedicated tool/web-freshness line is still being computed.
## Identity
Reframr is built by **OkeyMeta Ltd**. The Reframr line reframes language intelligence around recurrent memory, computed weights, and evidence from data. OkeyMeta Ltd was founded in 2022. The founder and CEO is **Okechukwu Goodnews Nwaozor**.
## Architecture Snapshot
| Property | Reframr-RFM-v1-Base |
| --- | --- |
| Family | Reframr / Recurrent Flow Memory |
| Organization | OkeyMeta Ltd |
| Checkpoint kind | `reframr-analytical` |
| Attention stack | None |
| Transformer layers | None |
| Tokenizer | FrameToken |
| Weight file | `model.safetensors` |
| Runtime | CPU-first Reframr Python runtime |
| Embedding dim | 96 |
| State dim | 48 |
| State width | 576 |
| Output vocab rows | 2,793 |
| Tokenizer vocab size | 3,741 |
## Intended Use
This checkpoint is intended for public testing of the Reframr runtime, open-ended generation experiments, system-instruction experiments, story generation, safety behavior, identity prompts, and CPU-first research into non-Transformer language modeling.
It is a base checkpoint, not a medical, legal, financial, or safety-critical authority. For fresh factual questions, connect a retrieval or web-search tool in the next tool-aware Reframr line rather than relying on static checkpoint knowledge alone.
## Release Note
This release is the public v1 base checkpoint. Internally, it comes from the v95 tracked compute run; publicly, it begins the Reframr-RFM v1 line. The next production line is being computed with broader data, tool-use supervision, web-search protocol tokens, and larger generalization probes. The goal is simple: make Reframr a serious, CPU-first, non-Transformer model family that learns from data rather than from hardcoded responses.
## Ownership
Copyright OkeyMeta Ltd. All rights reserved unless a separate license is supplied by OkeyMeta Ltd.
|