| # ROSA+ |
| **ROSA+**: RWKV's ROSA implementation with fallback statistical predictor |
| <hr/> |
| <img width="700" alt="image" src="https://github.com/user-attachments/assets/606a4d61-87b8-4bfc-ac27-564528042605" /> |
| <hr/> |
|
|
| ## What is ROSA+? |
| **ROSA+** is an extension of the **statistical next-token predictor** [proposed by BlinkDL](https://x.com/BlinkDL_AI/status/1976912771985146184) in extending the RWKV language model. It provides an intuitive Python interface as well as a fallback Witten–Bell predictor for unknown sequences. |
|
|
| ## Example Usage in Python |
| The implementation is self-contained in `rosaplus.py`. You can download the repository and use it from there. |
|
|
| ```python |
| # example.py |
| from rosaplus import ROSAPlus |
| import requests |
| |
| # Train on Shakespare |
| url = "https://raw.githubusercontent.com/karpathy/char-rnn/master/data/tinyshakespeare/input.txt" |
| |
| # Download the text |
| response = requests.get(url) |
| text = response.text |
| print("Downloaded text.") |
| |
| # Initialize model |
| m = ROSAPlus(max_order=1048576, use_eot=False, seed=0) |
| m.train_example(text) # Train ROSA |
| m.build_lm() # Train fallback predictor |
| |
| # Prompting |
| prompt = "ROMEO:" # Novel text |
| max_tokens = 256 |
| |
| # Eval mode |
| print(prompt + m.generate(prompt, steps=max_tokens)) |
| |
| # Saving model |
| m.save("rosa-model.json") |
| m2 = ROSAPlus.load("rosa-model.json") # Loading model |
| ``` |
|
|
| **Output:** (verbatim) |
| ``` |
| ROMEO: |
| In faith, I will. Let me peruse this face. |
| Mercutio's kinsman, noble County Paris! |
| What said my man, when my betossed soul |
| Did not attend him as we rode? I think |
| ... |
| ``` |
|
|
| ## Novel Text Generation |
| ROSA+ can also be used to generate novel sequences that **do not show up in the training dataset.** You can enable this by **always** using the fallback predictor. It often leads to coherent, surprising results. |
| ```python |
| # add always_fallback=True to the example |
| print(prompt + m.generate(prompt, steps=max_tokens, always_fallback=True)) |
| ``` |
|
|
| **Output:** (novel) |
| ``` |
| ROMEO: |
| The exchange of joy |
| That only Warwick's daughter. |
| |
| CLARENCE: |
| To whom, my lord; the foe vantage. |
| But make you read no other, and look'd deadly that name remains; |
| The cruelty and envy of the people, |
| Permitted by our faces |
| For man or master; then it for some |
| ``` |
|
|
| As you can see, these arrangement of sentences do not show up in the dataset (CTRL+F). Rather, ROSA+ intelligently splices and pulls together the features from ROSA to perform next-character prediction. |
|
|
| For any given prefix, you can also get the probability distribution for the next token: |
| ```python |
| # Eval mode |
| print(m.get_dist("ROMEO:\nOh, how could yo")) |
| ``` |
|
|
| Output: |
| ``` |
| {'u': 0.9999989177710094, 'n': 5.442332067424175e-07, 'k': 5.379892385443467e-07, 'r': 6.0439900862193395e-12, ' ... |
| ``` |
|
|
| As you can see, ROSA+ is extremely confident that 'u' is the next token (and it is correct!) |
|
|
| ## RWKV Support |
| This is just a standalone example of ROSA and does not provide RWKV integration. You will have to go to the RWKV Discord or ask the main maintainer (BlinkDL) for assistance in this regard. |
|
|
| ## Extensions |
| ROSA+ extends ROSA by: |
|
|
| - Allowing training and sampling on individual sequences, similar to a LLM |
| - Utilizing a (coherent) fallback Witten–Bell based predictor for when ROSA is unsure of the next token. |
|
|
| This makes it extremely fast, since ROSA is used for 99% of the predictions and the fallback only occurs for novel sequences. |
|
|
| **Tokenization:** The default tokenization is character-based (I will add support for new tokenizers coming soon.) |
|
|
| ## Notes |
|
|
| If you install [orjson](https://github.com/ijl/orjson), it will use it automatically and lead to far faster import/export speed. |
| Docs coming soon. |
|
|
| ## Issues with ROSA+ |
|
|
| ROSA+ is entirely statistical-based -- it extends upon the ROSA predictor proposed by BlinkDL, then provides a probability predictor as a fallback. However, this means it only has a **database-like** understanding of text -- it can **stitch together multiple sentences** and **demonstrate grammar**, but it lacks the same context understanding as an NN (RWKV, Transformer etc.) |
|
|
| For instance, when trained on Shakespeare, and with `always_fallback=True` (forcing novel predictions), it generates text that "looks right", but switches between characters every stanza. |
|
|
| ``` |
| COMINIUS: |
| Well, one nail; |
| Right noble is thy mercy dried their watches of chance and thy lord's false love; |
| For both of you are birds of selfsame feather. |
| |
| KING EDWARD IV: |
| Peace, wilful boy, or I will charm your tongue. |
| |
| CLARENCE: |
| Unhappy fortune! by my troth, I looked upon his faith iron cook you, sir, he bid me knocks; ha! let me be unrolled and said 'The better for our purpose.' |
| |
| KING RICHARD III: |
| So proud the name of Henry with your holy look'd on me, |
| And wouldst do not break your oaths; for of that sin |
| May deny her aiding have these nothing here. |
| |
| AUTOLYCUS: |
| I hope so, sir; for I have about me manner doth accuse my husband, I |
| ... |
| ``` |
|
|
| A ChatGPT analysis of ROSA+'s lines uncovers some insight: |
| ``` |
| Short answer: it’s Shakespeare-flavored, not Shakespearean. It reads like a collage of misquoted or remixed lines, with scrambled idioms, mixed plays (Juliet/Romeo with Buckingham and Gaunt), and meter/grammar that don’t line up with blank verse. |
| |
| Quick notes: |
| |
| * “Now, by Saint Peter’s Church…” and “I have forgot why I did call thee back” echo *Romeo and Juliet*, but they’re spliced into new contexts. |
| * “The world goes his bonnet to an oystery” mangles Pistol’s “The world’s mine oyster.” |
| * Shifts between **you/thee/thou/thy** are inconsistent (use *thou* as subject, *thee* as object, *thy/thine* as possessives). |
| * Many lines don’t scan as iambic pentameter (10 syllables, mostly unstressed–stressed). |
| ``` |
|
|
| A true NN-based model would outperform a standalone ROSA+ implementation because of the understanding of actual context. While ROSA+ has impressive surface-level understanding, it lacks deeper, low level meaning expressed by NNs. |
|
|
|
|
| ### Interesting occurences |
| You can view all the samples in the `samples` directory -- interestingly, in `sample_default_output.txt`, the model falls into an attractor state, repeating itself every ~3k lines, halfway through. However, in `sample_novel_output.txt`, you can spot some very novel sentences: |
| ``` |
| LADY ANNE: |
| Well, well, peace be with you, sir, he bid me know the points o' the dead |
| May walk again: if such thing but what I am, |
| I would wish it gone, |
| ``` |
|
|
| The phrases `Well, well, peace be with you` and `I would wish it gone` never show up in the training data. |
|
|
| ## Use cases |
| - Autocorrect / word prediction |
| - Translation (possibly) |
| - Features for a lower level model |
| - Generating surface-level text that fools detectors |
|
|
| ## Improving ROSA+ |
| One may be able to create a coherent language model simply by feeding ROSA+ embeddings into a GRU. Since ROSA+ captures the immediate surface-level features of text, a sufficient neural network may be able to operate on these embeddings and alter the distribution for more fine-grained understanding. |
|
|
| ## Downsides of statistical LMs |
| Unless statistical LMs incorporate some kind of **statistical attention mechanism** (which is possible!) they will never be able to grasp a high-level understanding of text as do humans and LMs. A statistical LM is unable to **copy data / tokens from one place to another**, **operate on a continous state**, **blend together tokens across different spaces**, perform **few-shot learning** (needs neural mechanism!) or **transfer learning** (no state vectors!). Therefore, their purpose remains limited to grasping **surface-level** features of text, like syntax, or general structure. |
|
|
| Google pushed to make their translation software (which in the 2010s, was n-gram based) the best at the time, but even **LSTMs** (which were invented way before Transformers) managed to outperform them. |
|
|
| Do not let this discourage you though. It may be practical to incorporate some kind of **continous state vector / representation** within a statistical model, making it **drastically more efficient than LLMs** while preserving all the benefits of **NN-based models.** This is an active field of research at **Bellevue College ML** (BCML) -- and if pioneered, could result in language models thousands of times more efficient. Don't let an article discourage you. |
|
|