Bochkov commited on
Commit
6aab3eb
·
verified ·
1 Parent(s): 14ae094

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -1
README.md CHANGED
@@ -46,4 +46,23 @@ emb_path = hf_hub_download(
46
  filename="normalized_embeddings_weights.pt"
47
  )
48
 
49
- embeddings = torch.load(emb_path)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
46
  filename="normalized_embeddings_weights.pt"
47
  )
48
 
49
+ embeddings = torch.load(emb_path)
50
+ ```
51
+
52
+ ## 🧑‍🔬 Citation & Concept
53
+
54
+ If you use this model or the underlying concepts in your research, please cite our work:
55
+
56
+ ```
57
+ @misc{bochkov2025emergentsemanticstokenembeddings,
58
+ title={Emergent Semantics Beyond Token Embeddings: Transformer LMs with Frozen Visual Unicode Representations},
59
+ author={A. Bochkov},
60
+ year={2025},
61
+ eprint={2507.04886},
62
+ archivePrefix={arXiv},
63
+ primaryClass={cs.CL},
64
+ url={https://arxiv.org/abs/2507.04886},
65
+ }
66
+ ```
67
+
68
+ This work demonstrates that transformer blocks, not token embeddings, carry the semantic burden in LLMs — a step toward modular, fusable, multilingual LMs.