Update README.md
Browse files
README.md
CHANGED
|
@@ -1,3 +1,37 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: apache-2.0
|
| 3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
tags:
|
| 4 |
+
- codon
|
| 5 |
+
- cds
|
| 6 |
+
- CDS
|
| 7 |
+
- mRNA
|
| 8 |
+
- RNA
|
| 9 |
+
- Codon
|
| 10 |
+
---
|
| 11 |
+
# Dataset Card for Dataset Name
|
| 12 |
+
|
| 13 |
+
<!-- Provide a quick summary of the dataset. -->
|
| 14 |
+
|
| 15 |
+
This is the dataset that was used to train SynCodonLM.
|
| 16 |
+
It is made up of roughly 66 million CDS sequences (without introns), from roughyl 35,000 species.
|
| 17 |
+
The github of our model SynCodonLM can be [found here].(https://github.com/Boehringer-Ingelheim/SynCodonLM)
|
| 18 |
+
The Hugging Face of our learned model weights can be [found here] (https://huggingface.co/jheuschkel/SynCodonLM).
|
| 19 |
+
|
| 20 |
+
|
| 21 |
+
|
| 22 |
+
## Citation
|
| 23 |
+
If you use this work, please cite:
|
| 24 |
+
```bibtex
|
| 25 |
+
@article {Heuschkel2025.08.19.671089,
|
| 26 |
+
author = {Heuschkel, James and Kingsley, Laura and Pefaur, Noah and Nixon, Andrew and Cramer, Steven},
|
| 27 |
+
title = {Advancing Codon Language Modeling with Synonymous Codon Constrained Masking},
|
| 28 |
+
elocation-id = {2025.08.19.671089},
|
| 29 |
+
year = {2025},
|
| 30 |
+
doi = {10.1101/2025.08.19.671089},
|
| 31 |
+
publisher = {Cold Spring Harbor Laboratory},
|
| 32 |
+
abstract = {Codon language models offer a promising framework for modeling protein-coding DNA sequences, yet current approaches often conflate codon usage with amino acid semantics, limiting their ability to capture DNA-level biology. We introduce SynCodonLM, a codon language model that enforces a biologically grounded constraint: masked codons are only predicted from synonymous options, guided by the known protein sequence. This design disentangles codon-level from protein-level semantics, enabling the model to learn nucleotide-specific patterns. The constraint is implemented by masking non-synonymous codons from the prediction space prior to softmax. Unlike existing models, which cluster codons by amino acid identity, SynCodonLM clusters by nucleotide properties, revealing structure aligned with DNA-level biology. Furthermore, SynCodonLM outperforms existing models on 6 of 7 benchmarks sensitive to DNA-level features, including mRNA and protein expression. Our approach advances domain-specific representation learning and opens avenues for sequence design in synthetic biology, as well as deeper insights into diverse bioprocesses.Competing Interest StatementThe authors have declared no competing interest.},
|
| 33 |
+
URL = {https://www.biorxiv.org/content/early/2025/08/24/2025.08.19.671089},
|
| 34 |
+
eprint = {https://www.biorxiv.org/content/early/2025/08/24/2025.08.19.671089.full.pdf},
|
| 35 |
+
journal = {bioRxiv}
|
| 36 |
+
}
|
| 37 |
+
```
|