yanjiehuang commited on
Commit
0a68408
·
verified ·
1 Parent(s): 7b65fce

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +9 -4
README.md CHANGED
@@ -16,7 +16,7 @@ pipeline_tag: text-generation
16
  # EVA: Evolutionary Versatile Architect
17
 
18
  EVA is a generative foundation model for universal RNA modeling and design,
19
- trained on **RNAVerse v1** — a curated atlas of **114 million full-length RNA
20
  sequences** spanning all domains of life.
21
 
22
  ## Model Description
@@ -24,12 +24,17 @@ sequences** spanning all domains of life.
24
  | Property | Details |
25
  |----------|---------|
26
  | Architecture | Decoder-only Transformer + Mixture-of-Experts (MoE) |
27
- | Parameters | 1.4B (also available: 21M, 167M, 470M) |
28
  | Context Window | 8,192 tokens |
29
- | Training Data | 114M full-length RNA sequences (RNAVerse v1) |
30
- | Training Objectives | Causal LM (CLM) + Generalized LM (GLM), 2:1 ratio |
31
  | Conditioning | RNA type tags + taxonomic lineage tags |
32
 
 
 
 
 
 
33
  For instructions, details, and examples, please refer to our
34
  [technical report](https://arxiv.org/abs/XXXX.XXXXX) and
35
  [GitHub repository](https://github.com/kevinhyj/EVA1).
 
16
  # EVA: Evolutionary Versatile Architect
17
 
18
  EVA is a generative foundation model for universal RNA modeling and design,
19
+ trained on **OpenRNA v1** — a curated atlas of **114 million full-length RNA
20
  sequences** spanning all domains of life.
21
 
22
  ## Model Description
 
24
  | Property | Details |
25
  |----------|---------|
26
  | Architecture | Decoder-only Transformer + Mixture-of-Experts (MoE) |
27
+ | Parameters | 1.4B (also available: 21M, 145M, 437M) |
28
  | Context Window | 8,192 tokens |
29
+ | Training Data | 114M full-length RNA sequences (OpenRNA v1) |
30
+ | Training Objectives | Causal LM (CLM) + Generalized LM (GLM) |
31
  | Conditioning | RNA type tags + taxonomic lineage tags |
32
 
33
+ > **Model Variants**
34
+ > - `EVA_21M`, `EVA_145M`, `EVA_437M`: trained with both CLM and GLM objectives, supporting both generation modes.
35
+ > - `EVA_1.4B_GLM`: the primary 1.4B model, trained with both CLM and GLM objectives.
36
+ > - `EVA_1.4B_CLM`: an additional 1.4B checkpoint trained exclusively with the CLM objective.
37
+
38
  For instructions, details, and examples, please refer to our
39
  [technical report](https://arxiv.org/abs/XXXX.XXXXX) and
40
  [GitHub repository](https://github.com/kevinhyj/EVA1).