Feature Extraction
Transformers
gpt2
Bochkov commited on
Commit
2fe5112
·
verified ·
1 Parent(s): 6de96d0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -0
README.md CHANGED
@@ -64,6 +64,16 @@ If you use this model or the underlying concepts in your research, please cite o
64
  primaryClass={cs.CL},
65
  url={https://arxiv.org/abs/2507.04886},
66
  }
 
 
 
 
 
 
 
 
 
 
67
  ```
68
 
69
  This work demonstrates that transformer blocks, not token embeddings, carry the semantic burden in LLMs — a step toward modular, fusable, multilingual LMs.
 
64
  primaryClass={cs.CL},
65
  url={https://arxiv.org/abs/2507.04886},
66
  }
67
+
68
+ @misc{bochkov2025growingtransformersmodularcomposition,
69
+ title={Growing Transformers: Modular Composition and Layer-wise Expansion on a Frozen Substrate},
70
+ author={A. Bochkov},
71
+ year={2025},
72
+ eprint={2507.07129},
73
+ archivePrefix={arXiv},
74
+ primaryClass={cs.LG},
75
+ url={https://arxiv.org/abs/2507.07129},
76
+ }
77
  ```
78
 
79
  This work demonstrates that transformer blocks, not token embeddings, carry the semantic burden in LLMs — a step toward modular, fusable, multilingual LMs.