Instructions to use ilbert/reflex-coder7b-jepa-riscv with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use ilbert/reflex-coder7b-jepa-riscv with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="ilbert/reflex-coder7b-jepa-riscv")# Load model directly from transformers import AutoModel model = AutoModel.from_pretrained("ilbert/reflex-coder7b-jepa-riscv", dtype="auto") - Notebooks
- Google Colab
- Kaggle
- Local Apps
- vLLM
How to use ilbert/reflex-coder7b-jepa-riscv with vLLM:
Install from pip and serve model
# Install vLLM from pip: pip install vllm # Start the vLLM server: vllm serve "ilbert/reflex-coder7b-jepa-riscv" # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:8000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "ilbert/reflex-coder7b-jepa-riscv", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }'Use Docker
docker model run hf.co/ilbert/reflex-coder7b-jepa-riscv
- SGLang
How to use ilbert/reflex-coder7b-jepa-riscv with SGLang:
Install from pip and serve model
# Install SGLang from pip: pip install sglang # Start the SGLang server: python3 -m sglang.launch_server \ --model-path "ilbert/reflex-coder7b-jepa-riscv" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "ilbert/reflex-coder7b-jepa-riscv", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }'Use Docker images
docker run --gpus all \ --shm-size 32g \ -p 30000:30000 \ -v ~/.cache/huggingface:/root/.cache/huggingface \ --env "HF_TOKEN=<secret>" \ --ipc=host \ lmsysorg/sglang:latest \ python3 -m sglang.launch_server \ --model-path "ilbert/reflex-coder7b-jepa-riscv" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "ilbert/reflex-coder7b-jepa-riscv", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }' - Docker Model Runner
How to use ilbert/reflex-coder7b-jepa-riscv with Docker Model Runner:
docker model run hf.co/ilbert/reflex-coder7b-jepa-riscv
| license: mit | |
| base_model: Qwen/Qwen2.5-Coder-7B-Instruct | |
| tags: | |
| - riscv | |
| - cross-attention | |
| - flamingo | |
| - grounded | |
| - jepa | |
| - rv32i | |
| - code-generation | |
| library_name: transformers | |
| pipeline_tag: text-generation | |
| # Reflex-Coder7B-JEPA-RISCV | |
| **A frozen `Qwen2.5-Coder-7B-Instruct` wired to a RISC-V CPU through Flamingo-style cross-attention. Emits one 32-bit RV32I instruction per cycle, conditioned on live machine state. The output head is a JEPA-style embedding predictor over a learned 691-row instruction codebook; nearest-neighbour decode gives free error-correction on individual predictions.** | |
| This repo contains the **adapter weights only** (~4.4 GB). The frozen backbone is pulled from `Qwen/Qwen2.5-Coder-7B-Instruct` at runtime. Total inference footprint: ~14 GB bf16 backbone + 4.4 GB adapters + activations. | |
| ## What it does | |
| Given a natural-language prompt (`"multiply 7 and 8"`, `"compute 5 factorial"`, `"say hi"`), Reflex drives a Unicorn-backed RV32I emulator instruction by instruction. Each cycle: | |
| 1. Read live CPU state (32 registers, PC, memory windows around PC and SP). | |
| 2. Encode as 65 K/V tokens. | |
| 3. Run the frozen backbone forward over the prompt; cross-attn adapters fuse state K/V into hidden states every 4 layers. | |
| 4. Last-token pool → MLP → 256-d embedding. | |
| 5. Cosine nearest-neighbour against a 691-row instruction codebook → a real 32-bit RV32I word. | |
| 6. Write the word at PC in Unicorn, step one cycle, loop. | |
| ## Base model | |
| [`Qwen/Qwen2.5-Coder-7B-Instruct`](https://huggingface.co/Qwen/Qwen2.5-Coder-7B-Instruct) — frozen, bf16, untouched. | |
| ## Training | |
| - **Corpus**: 80,396 `(prompt, program)` pairs across 56 RV32I program families (arithmetic, loops, comparisons, memory ops, display writes). Every program verified by running it end-to-end through Unicorn before training (zero rejects). | |
| - **Flattened cycle pool**: ~1.06 M `(state, next_instruction)` pairs, subsampled to ~173 k balanced across families. | |
| - **Objective**: InfoNCE (temperature Ï„ = 0.07) over the full 691-row instruction codebook. The codebook rows train jointly with the controller. | |
| - **Optimizer**: AdamW (weight decay 0.01), cosine LR `1e-4 → 1e-6` over 15 000 steps, batch 16. | |
| - **Hardware**: single A100 80 GB (~4 h) or L40S 48 GB (batch 32, ~5 h). | |
| ## Results (41-task eval) | |
| | section | pass | | |
| |---|---| | |
| | in-distribution (8) | **7 / 8** | | |
| | out-of-distribution (10) | **9 / 10** | | |
| | display strings (4) | 1 / 4 | | |
| | novel zero-shot (9) | **7 / 9** | | |
| | consistency: factorial 5 × 10 | **10 / 10** | | |
| | **total** | **34 / 41 (83 %)** | | |
| Highlights: | |
| - **`popcount(255) = 8` in 199 consecutive correct RISC-V instructions** — emergent bit-counting loop the model was never trained on. | |
| - **Factorial 5 × 10 = 120, deterministic** — every run emits exactly 91 ops and lands on the right answer. | |
| - Zero-shot `multiply 7×8`, `power 2^5`, `min(7,3,9)`, `abs(-5)`, `count up 1..5` all pass. | |
| Per-step top-1 instruction accuracy on 500 random held-out cycles: **96.0 %**. All `BRANCH`, `R-type`, `LOAD`, `STORE`, `JAL`, `JALR` predictions are 100 %. **Every top-1 miss is same-opcode** — never an opcode flip. | |
| ## Usage | |
| ```python | |
| from reflex.demo import load, run_grounded | |
| model, tok, cfg = load("reflex.pt", device="cuda") | |
| cpu, emitted, halted, err = run_grounded( | |
| model, tok, "multiply 7 and 8", device="cuda", max_cycles=200, | |
| ) | |
| print(f"halted={halted} mem[0x5000]={cpu.mem_word(0x5000)}") | |
| # halted=True mem[0x5000]=56 | |
| ``` | |
| Or, interactively: | |
| ```bash | |
| uv run demo --checkpoint reflex.pt | |
| ``` | |
| ## Installation | |
| ```bash | |
| git clone https://github.com/ilbertt/reflex | |
| cd reflex | |
| uv sync | |
| huggingface-cli download ilbertt/reflex-coder7b-jepa-riscv reflex.pt --local-dir . | |
| ``` | |
| On first run, HuggingFace will automatically fetch `Qwen2.5-Coder-7B-Instruct` (~15 GB). | |
| ## Limitations | |
| - **Display byte-constants are unreliable.** The model picks ASCII neighbours: `show 42` writes `'·0'` instead of `'42'`; `print hello` writes `'hell·'`. These are same-opcode ±1-immediate misses, not opcode flips. | |
| - **Uncommon-literal arithmetic drifts.** `add 100+200` sometimes halts with 120; `double 100` → 0 in some seeds. Failures concentrate on `ADDI`/`LUI` with rare immediate values. | |
| - **Closed action space.** The codebook has exactly 691 rows — instructions never seen in training have no row and cannot be emitted. Ample for the 56 program families trained on; bounds generalisation to genuinely unseen opcodes. | |
| - **No domain-knowledge transfer.** Prompts like `"x5 is fever, display SICK"` fail. The adapters only route the backbone's prior through for program-shaped prompts seen in training. | |
| - **RV32I base ISA only.** No M, Zbb, F extensions. | |
| ## Files | |
| - `reflex.pt` — adapter weights, state encoder, cross-attn adapters, embedding head, 691-row instruction codebook, instruction-word buffer, and config dict (`backbone_id`, `hidden`, `inject_every`, `adapter_mlp_ratio`, `max_instr_tokens`, `embed_dim`, `num_instrs`, `chat_template`, `context_prefix`). | |