Update README.md
Browse files
README.md
CHANGED
|
@@ -46,4 +46,23 @@ emb_path = hf_hub_download(
|
|
| 46 |
filename="normalized_embeddings_weights.pt"
|
| 47 |
)
|
| 48 |
|
| 49 |
-
embeddings = torch.load(emb_path)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 46 |
filename="normalized_embeddings_weights.pt"
|
| 47 |
)
|
| 48 |
|
| 49 |
+
embeddings = torch.load(emb_path)
|
| 50 |
+
```
|
| 51 |
+
|
| 52 |
+
## 🧑🔬 Citation & Concept
|
| 53 |
+
|
| 54 |
+
If you use this model or the underlying concepts in your research, please cite our work:
|
| 55 |
+
|
| 56 |
+
```
|
| 57 |
+
@misc{bochkov2025emergentsemanticstokenembeddings,
|
| 58 |
+
title={Emergent Semantics Beyond Token Embeddings: Transformer LMs with Frozen Visual Unicode Representations},
|
| 59 |
+
author={A. Bochkov},
|
| 60 |
+
year={2025},
|
| 61 |
+
eprint={2507.04886},
|
| 62 |
+
archivePrefix={arXiv},
|
| 63 |
+
primaryClass={cs.CL},
|
| 64 |
+
url={https://arxiv.org/abs/2507.04886},
|
| 65 |
+
}
|
| 66 |
+
```
|
| 67 |
+
|
| 68 |
+
This work demonstrates that transformer blocks, not token embeddings, carry the semantic burden in LLMs — a step toward modular, fusable, multilingual LMs.
|