Add model card and metadata for MemoBrain-14B

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +77 -0
README.md ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ library_name: transformers
4
+ pipeline_tag: text-generation
5
+ ---
6
+
7
+ # MemoBrain-14B: Executive Memory as an Agentic Brain for Reasoning
8
+
9
+ MemoBrain-14B is an executive memory model designed for tool-augmented agents. It constructs a dependency-aware memory over reasoning steps, capturing salient intermediate states and their logical relations to sustain coherent, goal-directed reasoning over long horizons.
10
+
11
+ - **Paper:** [MemoBrain: Executive Memory as an Agentic Brain for Reasoning](https://huggingface.co/papers/2601.08079)
12
+ - **Repository:** [https://github.com/qhjqhj00/MemoBrain](https://github.com/qhjqhj00/MemoBrain)
13
+
14
+ ## Overview
15
+ Complex reasoning in tool-augmented agent frameworks is often long-horizon, causing reasoning traces to strain the working context of LLMs. MemoBrain operates as a co-pilot alongside the reasoning agent, managing the working context by:
16
+ - **Pruning** invalid steps.
17
+ - **Folding** completed sub-trajectories into compact summaries.
18
+ - **Preserving** a high-salience reasoning backbone under a fixed context budget.
19
+
20
+ MemoBrain-14B is based on the Qwen3 architecture and has been specifically fine-tuned for these memory operations.
21
+
22
+ ## Quick Start
23
+
24
+ ### Deployment with vLLM
25
+ We recommend deploying the model using [vLLM](https://github.com/vllm-project/vllm) for high-performance inference:
26
+
27
+ ```bash
28
+ pip install vllm
29
+ vllm serve TommyChien/MemoBrain-14B --port 8002
30
+ ```
31
+
32
+ ### Python Usage
33
+ Once the model is served, you can interact with it using the `memobrain` package from the [official repository](https://github.com/qhjqhj00/MemoBrain).
34
+
35
+ ```python
36
+ import asyncio
37
+ from memobrain import MemoBrain
38
+
39
+ async def main():
40
+ # Step 1: Initialize MemoBrain
41
+ memory = MemoBrain(
42
+ api_key="EMPTY", # vLLM doesn't require API key
43
+ base_url="http://localhost:8002/v1",
44
+ model_name="TommyChien/MemoBrain-14B"
45
+ )
46
+
47
+ # Step 2: Initialize memory with your task
48
+ memory.init_memory("Solve a complex research problem")
49
+
50
+ # Step 3: Memorize conversation interactions
51
+ # The recommended unit is an episode: thinking → tool call → tool response
52
+ await memory.memorize([
53
+ {"role": "assistant", "content": "I need to search for information about Paris..."},
54
+ {"role": "user", "content": "Search results: Paris is the capital of France..."}
55
+ ])
56
+
57
+ # Step 4: Optimize memory (flush invalid steps & fold completed sub-trajectories)
58
+ optimized_messages = await memory.recall()
59
+ print(f"Memory optimized: {len(optimized_messages)} messages")
60
+
61
+ asyncio.run(main())
62
+ ```
63
+
64
+ ## Experimental Results
65
+ MemoBrain-8B and 14B have demonstrated state-of-the-art performance on long-horizon reasoning benchmarks such as GAIA and WebWalker, showing significant improvements particularly on complex, multi-step tasks.
66
+
67
+ ## Citation
68
+ If you find MemoBrain useful for your research, please cite:
69
+
70
+ ```bibtex
71
+ @article{memobrain2026,
72
+ title={MemoBrain: Executive Memory as an Agentic Brain for Reasoning},
73
+ author={Hongjin Qian, Zhao Cao, Zheng Liu},
74
+ journal={arXiv preprint arXiv:2601.08079},
75
+ year={2026}
76
+ }
77
+ ```