FutureLivingLab commited on
Commit
08b5406
·
verified ·
1 Parent(s): 4ae1687

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +106 -0
README.md ADDED
@@ -0,0 +1,106 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ pipeline_tag: text-generation
4
+ ---
5
+
6
+ # ROME-30B-A3B (Coming Soon)
7
+
8
+ <a href="https://arxiv.org/pdf/2512.24873" target="_blank">
9
+ 🔗 <strong>Technical Report</strong><br/>
10
+ <img alt="Paper" src="https://img.shields.io/badge/Paper-arXiv%3A2512.24873-red"/>
11
+ </a>
12
+
13
+
14
+ ---
15
+
16
+ ## 📢 Note: Coming Soon!
17
+
18
+ **ROME (ROME is Obviously an Agentic ModEl)** will be officially released soon.
19
+ The project is currently under final review and preparation. Model weights will be made publicly available shortly. *(The iFlow CLI has already been released.)* Stay tuned!
20
+
21
+ <img src="https://rlhf.oss-cn-hangzhou.aliyuncs.com/iFLOW-ROME/performance.png" width="600"/>
22
+
23
+
24
+ ---
25
+
26
+
27
+
28
+ ## Highlights
29
+
30
+ **ROME** is an open-source **agentic foundation model** incubated within the **ALE (Agentic Learning Ecosystem)**.
31
+
32
+ Rather than scaling performance purely by increasing parameter count, ROME achieves *parameter-scale–crossing agentic performance* through **full-stack infrastructure and algorithmic optimization**.
33
+
34
+ <img src="https://rlhf.oss-cn-hangzhou.aliyuncs.com/iFLOW-ROME/ALE.PNG" width="600"/>
35
+
36
+
37
+ ### 🔧 ALE Full-Stack Infrastructure
38
+ - **ROLL** – Large-scale reinforcement learning optimization engine
39
+
40
+ - **ROCK** – Secure sandbox and environment orchestration for agent execution
41
+
42
+ - **iFlow CLI** – Unified agent framework and developer interface
43
+
44
+
45
+
46
+ ### 🧠 IPA Policy Optimization Algorithm
47
+ - Introduces **Interaction-Perceptive Agentic Policy Optimization (IPA)**
48
+ - Performs credit assignment at the level of **Semantic Interaction Chunks**
49
+ - Significantly improves **training stability** and **success rates** on **long-horizon tasks**
50
+
51
+
52
+
53
+ ### 🚀 Strong Agentic Performance
54
+ - Despite being a **mid-sized model** (30B MoE with 3B active parameters), ROME outperforms same-scale models on standard agent benchmarks:
55
+ - **Terminal-Bench 2.0**: 24.72%
56
+ - **SWE-bench Verified**: 57.40%
57
+
58
+ - Performance is competitive with, and in some cases comparable to, models exceeding **100B parameters**
59
+
60
+
61
+
62
+ ### 🔒 Production-Grade Safety
63
+ - Designed for autonomous agent execution in real environments
64
+ - Rigorously aligned and red-teamed against risks such as:
65
+ - Unauthorized access
66
+ - Illegal or unsafe tool invocation
67
+ - Built with **deployment-grade safety guarantees** in mind
68
+
69
+ ---
70
+
71
+
72
+
73
+ ## Performance (Preview)
74
+
75
+ ### Terminal-Based Benchmarks
76
+
77
+ | **Model** | **Terminal-Bench 2.0** | **SWE-bench Verified** |
78
+ | ---------------------------- | ---------------------- | ---------------------- |
79
+ | Qwen3-Coder-30B-A3B-Instruct | 13.48% | 46.33% |
80
+ | **ROME (30B-A3B)** | **24.72%** | **57.40%** |
81
+ | GPT-OSS-120B | 21.12% | 43.93% |
82
+ | GLM-4.5 Air (106B) | 17.30% | 56.20% |
83
+
84
+ > See the technical report for full experimental details.
85
+
86
+ ---
87
+
88
+ ## Best Practices
89
+
90
+ *(Code examples and usage guidelines will be added after the model release.)*
91
+
92
+ ---
93
+
94
+
95
+
96
+ ## Citation
97
+
98
+ If you find our work useful, please consider citing:
99
+
100
+ ```bibtex
101
+ @article{rome2025ale,
102
+ title={Let It Flow: Agentic Crafting on Rock and Roll - Building the ROME Model within an Open Agentic Learning Ecosystem},
103
+ author={ROME & ALE Team},
104
+ journal={arXiv preprint arXiv:2512.24873},
105
+ year={2025}
106
+ }