Update README.md
Browse files
README.md
CHANGED
|
@@ -19,9 +19,63 @@ datasets:
|
|
| 19 |
|
| 20 |
# BinaryLLM (HF export)
|
| 21 |
|
| 22 |
-
Tokenizer-free / base-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 23 |
|
| 24 |
-
## Load
|
| 25 |
```python
|
| 26 |
from transformers import AutoModelForCausalLM
|
| 27 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 19 |
|
| 20 |
# BinaryLLM (HF export)
|
| 21 |
|
| 22 |
+
Tokenizer-free / base-2 (vocab_size=4) model export.
|
| 23 |
+
|
| 24 |
+
This repo uses `trust_remote_code=True` because it ships a custom `modeling_*.py` / `configuration_*.py`.
|
| 25 |
+
|
| 26 |
+
## Load (Python)
|
| 27 |
|
|
|
|
| 28 |
```python
|
| 29 |
from transformers import AutoModelForCausalLM
|
| 30 |
+
|
| 31 |
+
m = AutoModelForCausalLM.from_pretrained(
|
| 32 |
+
"PhysiQuanty/B-LLM-Test",
|
| 33 |
+
trust_remote_code=True,
|
| 34 |
+
)
|
| 35 |
+
m.eval()
|
| 36 |
+
````
|
| 37 |
+
|
| 38 |
+
## Inference (CLI)
|
| 39 |
+
|
| 40 |
+
This repo includes a minimal inference script that:
|
| 41 |
+
|
| 42 |
+
* encodes the prompt to base-2 bits (UTF-8, MSB→LSB),
|
| 43 |
+
* runs a manual token-by-token loop (no `generate`),
|
| 44 |
+
* decodes the generated bits back to text (best-effort strict decode).
|
| 45 |
+
|
| 46 |
+
### Command
|
| 47 |
+
|
| 48 |
+
```bash
|
| 49 |
+
python3 inference.py --repo "PhysiQuanty/B-LLM-Test" --prompt "bonjour" --print_ids
|
| 50 |
+
```
|
| 51 |
+
|
| 52 |
+
### Example output
|
| 53 |
+
|
| 54 |
+
```text
|
| 55 |
+
[Seed] 857729874
|
| 56 |
+
[Device] cuda
|
| 57 |
+
[+] IDS = [2, 0, 1, 1, 0, 0, 0, 1, 0, 0, 1, 1, 0, 1, 1, 1, 1, 0, 1, 1, 0, 1, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 1, 1, 0, 1, 1, 1, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 3, 2]
|
| 58 |
+
|
| 59 |
+
[Prompt]
|
| 60 |
+
bonjour
|
| 61 |
+
|
| 62 |
+
[Prompt IDs] len=59 | BOS=2 EOS=3
|
| 63 |
+
|
| 64 |
+
[Output]
|
| 65 |
+
|
| 66 |
+
[Final Output]
|
| 67 |
+
|
| 68 |
+
L'évaluation suivante exprime quel sentiment ?
|
| 69 |
+
EvaluationSeuresser fragileuses et mais mais l'autr
|
| 70 |
+
|
| 71 |
+
[Generated IDs]
|
| 72 |
+
|
| 73 |
+
[0, 1, 0, 0, 1, 1, 0, 0, 0, 0, 1, 0, 0, 1, 1, 1, 1, 1, 0, 0, 0, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 1, 0, 1, 1, 0, 0, 1, 1, 0, 0, 0, 0, 1, 0, 1, 1, 0, 1, 1, 0, 0, 0, 1, 1, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 0, 0, 1, 0, 1, 1, 1, 0, 1, 0, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 1, 1, 1, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 1, 1, 1, 0, 0, 1, 1, 0, 1, 1, 1, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 1, 0, 1, 1, 0, 0, 1, 1, 0, 0, 0, 0, 1, 0, 1, 1, 0, 1, 1, 1, 0, 0, 1, 1, 1, 0, 1, 0, 0, 0, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 0, 0, 0, 0, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 1, 1, 0, 0, 0, 0, 1, 1, 1, 0, 0, 0, 0, 0, 1, 1, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 0, 0, 0, 0, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 1, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 1, 1, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 1, 1, 1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 1, 1, 1, 0, 0, 1, 1, 1, 0, 1, 0, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 1, 1, 1, 0, 0, 1, 1, 1, 0, 1, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 1, 0, 1, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 1, 1, 0, 0, 1, 1, 0, 0, 0, 0, 1, 0, 1, 1, 0, 1, 1, 0, 0, 0, 1, 1, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 0, 0, 1, 0, 1, 1, 1, 0, 1, 0, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 1, 1, 1, 0, 1, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 1, 0, 1, 1, 1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 1, 1, 0, 0, 1, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 0, 1, 1, 0, 0, 0, 0, 1, 0, 1, 1, 0, 0, 1, 1, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 1, 0, 0, 1, 0, 0, 0, 0, 0, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 1, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 1, 1, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 1, 0, 0, 1, 1, 0, 0, 1, 0, 0, 0, 0, 0, 0, 1, 1, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 1, 0, 0, 1, 1, 0, 0, 1, 0, 0, 0, 0, 0, 0, 1, 1, 0, 1, 1, 0, 0, 0, 0, 1, 0, 0, 1, 1, 1, 0, 1, 1, 0, 0, 0, 0, 1, 0, 1, 1, 1, 0, 1, 0, 1, 0, 1, 1, 1, 0, 1, 0, 0, 0, 1, 1, 1, 0, 0, 1, 0]
|
| 74 |
+
```
|
| 75 |
+
|
| 76 |
+
## Notes
|
| 77 |
+
|
| 78 |
+
* This model is **tokenizer-free**: the input prompt is encoded as base-2 bits (UTF-8 bytes → MSB→LSB).
|
| 79 |
+
* Some prompts may decode better than others depending on training distribution (e.g. frequent phrases).
|
| 80 |
+
|
| 81 |
+
```
|