Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
|
@@ -16,7 +16,7 @@ pipeline_tag: text-generation
|
|
| 16 |
# EVA: Evolutionary Versatile Architect
|
| 17 |
|
| 18 |
EVA is a generative foundation model for universal RNA modeling and design,
|
| 19 |
-
trained on **
|
| 20 |
sequences** spanning all domains of life.
|
| 21 |
|
| 22 |
## Model Description
|
|
@@ -24,12 +24,17 @@ sequences** spanning all domains of life.
|
|
| 24 |
| Property | Details |
|
| 25 |
|----------|---------|
|
| 26 |
| Architecture | Decoder-only Transformer + Mixture-of-Experts (MoE) |
|
| 27 |
-
| Parameters | 1.4B (also available: 21M,
|
| 28 |
| Context Window | 8,192 tokens |
|
| 29 |
-
| Training Data | 114M full-length RNA sequences (
|
| 30 |
-
| Training Objectives | Causal LM (CLM) + Generalized LM (GLM)
|
| 31 |
| Conditioning | RNA type tags + taxonomic lineage tags |
|
| 32 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
For instructions, details, and examples, please refer to our
|
| 34 |
[technical report](https://arxiv.org/abs/XXXX.XXXXX) and
|
| 35 |
[GitHub repository](https://github.com/kevinhyj/EVA1).
|
|
|
|
| 16 |
# EVA: Evolutionary Versatile Architect
|
| 17 |
|
| 18 |
EVA is a generative foundation model for universal RNA modeling and design,
|
| 19 |
+
trained on **OpenRNA v1** — a curated atlas of **114 million full-length RNA
|
| 20 |
sequences** spanning all domains of life.
|
| 21 |
|
| 22 |
## Model Description
|
|
|
|
| 24 |
| Property | Details |
|
| 25 |
|----------|---------|
|
| 26 |
| Architecture | Decoder-only Transformer + Mixture-of-Experts (MoE) |
|
| 27 |
+
| Parameters | 1.4B (also available: 21M, 145M, 437M) |
|
| 28 |
| Context Window | 8,192 tokens |
|
| 29 |
+
| Training Data | 114M full-length RNA sequences (OpenRNA v1) |
|
| 30 |
+
| Training Objectives | Causal LM (CLM) + Generalized LM (GLM) |
|
| 31 |
| Conditioning | RNA type tags + taxonomic lineage tags |
|
| 32 |
|
| 33 |
+
> **Model Variants**
|
| 34 |
+
> - `EVA_21M`, `EVA_145M`, `EVA_437M`: trained with both CLM and GLM objectives, supporting both generation modes.
|
| 35 |
+
> - `EVA_1.4B_GLM`: the primary 1.4B model, trained with both CLM and GLM objectives.
|
| 36 |
+
> - `EVA_1.4B_CLM`: an additional 1.4B checkpoint trained exclusively with the CLM objective.
|
| 37 |
+
|
| 38 |
For instructions, details, and examples, please refer to our
|
| 39 |
[technical report](https://arxiv.org/abs/XXXX.XXXXX) and
|
| 40 |
[GitHub repository](https://github.com/kevinhyj/EVA1).
|