afg1 commited on
Commit
c295d49
·
verified ·
1 Parent(s): 89f6cf2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +15 -8
README.md CHANGED
@@ -1,21 +1,27 @@
1
  ---
2
- tags:
3
- - generated_from_trainer
4
  model-index:
5
  - name: RNAMamba-14M
6
  results: []
 
 
 
 
7
  ---
8
 
9
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
10
- should probably proofread and complete it, then remove this comment. -->
11
 
12
  # RNAMamba-14M
13
 
14
- This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
15
 
16
- ## Model description
 
 
 
 
 
 
 
17
 
18
- More information needed
19
 
20
  ## Intended uses & limitations
21
 
@@ -26,6 +32,7 @@ More information needed
26
  More information needed
27
 
28
  ## Training procedure
 
29
 
30
  ### Training hyperparameters
31
 
@@ -43,4 +50,4 @@ The following hyperparameters were used during training:
43
  - Transformers 4.39.3
44
  - Pytorch 2.2.2+cu118
45
  - Datasets 2.18.0
46
- - Tokenizers 0.15.2
 
1
  ---
 
 
2
  model-index:
3
  - name: RNAMamba-14M
4
  results: []
5
+ license: apache-2.0
6
+ datasets:
7
+ - afg1/rnacentral_subset
8
+ pipeline_tag: fill-mask
9
  ---
10
 
 
 
11
 
12
  # RNAMamba-14M
13
 
14
+ This model is a small Mamba based model trained from scratch on 1.96 million sequences (1.56 billion bases) extracted from RNAcentral's active sequences FASTA file for release 24 (March 2024).
15
 
16
+ This is intended to be a sequence embedding model for downstream processing of ncRNA sequences.
17
+ It is trained with a masked language modelling objective, and a context size of 8,192 nucleotides.
18
+ The [dataset](https://huggingface.co/datasets/afg1/rnacentral_subset) has sequences ranging in length from 10 to 8192, so the model should be pretty good at handling sequences in that range.
19
+ This is a deliberately small model with only 14.1 million parameters (8 hidden layers, hidden dim 512, intermediate size 1024) such that it will run fast without a GPU. We may train something bigger if it looks like these embeddings are not good enough.
20
+
21
+
22
+ <!--## Model description
23
+ I'll fill this in later...
24
 
 
25
 
26
  ## Intended uses & limitations
27
 
 
32
  More information needed
33
 
34
  ## Training procedure
35
+ -->
36
 
37
  ### Training hyperparameters
38
 
 
50
  - Transformers 4.39.3
51
  - Pytorch 2.2.2+cu118
52
  - Datasets 2.18.0
53
+ - Tokenizers 0.15.2