Spaces:
Running
Running
Update README.md
Browse files
README.md
CHANGED
|
@@ -35,6 +35,9 @@ Our primary architecture - **Reactor** - is planned as the first _**awareness AG
|
|
| 35 |
connected to _Short-Term and Long-Term Memory_ (_Attention-based Memory System_) and _Receptors/Effectors_ systems for real-time reactive processing.
|
| 36 |
It will be able to constantly and autonomously learn from interactions in _Continouos Live Learning_ process.
|
| 37 |
|
|
|
|
|
|
|
|
|
|
| 38 |
## Reactive Language Models (RxLM)
|
| 39 |
While the **Reactor** is the main goal, it's extremely hard to achieve, as it's definitely the most advanced neural network ensemble ever.
|
| 40 |
|
|
@@ -44,7 +47,7 @@ That's why we designed simplified architectures, for incremental transformation
|
|
| 44 |
|
| 45 |
## RxLM vs LLM advantages
|
| 46 |
Processing single interactions in real-time by **Reactive Language Models** leads to **revolutional** improvements in inference speed/cost:
|
| 47 |
-
- LLM inference costs are increasing
|
| 48 |
- RxLM inference costs are linear, depending only on single interaction tokens (not accumulated) - each next interaction is `number of steps` times cheaper than for LLM
|
| 49 |
- same for inference speed - LLM has to process full history, while RxLM only single message (only first interaction could be slower because of encoder/memory attention overhead)
|
| 50 |
|
|
|
|
| 35 |
connected to _Short-Term and Long-Term Memory_ (_Attention-based Memory System_) and _Receptors/Effectors_ systems for real-time reactive processing.
|
| 36 |
It will be able to constantly and autonomously learn from interactions in _Continouos Live Learning_ process.
|
| 37 |
|
| 38 |
+
> Reactor architecture details were analysed by 30 state-of-the-art LLM/Reasoning models that rated it's potential
|
| 39 |
+
> to reach the AGI as ~4.35/5
|
| 40 |
+
|
| 41 |
## Reactive Language Models (RxLM)
|
| 42 |
While the **Reactor** is the main goal, it's extremely hard to achieve, as it's definitely the most advanced neural network ensemble ever.
|
| 43 |
|
|
|
|
| 47 |
|
| 48 |
## RxLM vs LLM advantages
|
| 49 |
Processing single interactions in real-time by **Reactive Language Models** leads to **revolutional** improvements in inference speed/cost:
|
| 50 |
+
- LLM inference costs are increasing quadratically with conversation length (accumulated for each next message), because of full dialog history processing
|
| 51 |
- RxLM inference costs are linear, depending only on single interaction tokens (not accumulated) - each next interaction is `number of steps` times cheaper than for LLM
|
| 52 |
- same for inference speed - LLM has to process full history, while RxLM only single message (only first interaction could be slower because of encoder/memory attention overhead)
|
| 53 |
|