Files changed (1) hide show
  1. README.md +77 -0
README.md ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ annotations_creators:
3
+ - no-annotation
4
+ language_creators:
5
+ - no-annotation
6
+ task_categories:
7
+ - text-classification
8
+ tags:
9
+ - genomics
10
+ - dna
11
+ - dnabert
12
+ - bioinformatics
13
+ - mouse-dna
14
+ - tokenized
15
+ source_datasets:
16
+ - davidcechak/Mouse_DNA_v0
17
+ language:
18
+ - en
19
+ license: other
20
+ license_name: unspecified
21
+ ---
22
+
23
+ # `Mouse_DNA_v0_DNABert6tokenized`
24
+
25
+ ## Dataset Description
26
+
27
+ The `davidcechak/Mouse_DNA_v0_DNABert6tokenized` is a processed version of the `davidcechak/Mouse_DNA_v0` dataset. It contains mouse DNA sequences that have been tokenized using a 6-mer approach, making it directly compatible with models like DNABert for classification and other downstream tasks.
28
+
29
+ This dataset can be used for comparative genomic analysis when used in conjunction with other tokenized datasets, such as the human DNA tokenized dataset (`simecek/Human_DNA_v0_DNABert6tokenized`). This allows for the training of classifiers that can distinguish between mouse and other species' DNA, providing a valuable resource for cross-species machine learning tasks in bioinformatics.
30
+
31
+ ## Dataset Structure
32
+
33
+ The dataset is available in the `parquet` format and is likely split into training and testing subsets, though this needs to be confirmed.
34
+
35
+ ### Data Fields
36
+
37
+ The dataset likely includes the following fields:
38
+ * **tokens**: A list of integers representing the 6-mer token IDs.
39
+ * **text**: The original DNA sequence string, consisting of the nucleotides `A`, `T`, `C`, and `G`.
40
+
41
+ ## Dataset Creation
42
+
43
+ ### Data Source
44
+
45
+ The base `Mouse_DNA_v0` dataset likely consists of DNA sequences from the mouse reference genome (*Mus musculus*).
46
+
47
+ ### Preprocessing and Tokenization
48
+
49
+ The raw sequences from the `Mouse_DNA_v0` dataset were processed using a 6-mer tokenization scheme:
50
+ 1. **Splitting**: Original DNA sequences were split into non-overlapping 6-mer tokens.
51
+ 2. **Mapping**: Each unique 6-mer was mapped to a unique integer ID to create a vocabulary.
52
+ 3. **Encoding**: The tokenized sequences were then represented as a list of these integer IDs.
53
+
54
+ ## Intended Uses
55
+
56
+ The dataset can be used for:
57
+ * **Comparative Genomics**: Comparing genomic features and training models to distinguish between species (e.g., mouse vs. human).
58
+ * **Genomic Classification**: Training and evaluating machine learning models on tasks like species identification from DNA sequences.
59
+ * **LLM Pre-training**: Providing a corpus for pre-training large language models on mouse DNA sequences, which can then be fine-tuned for more specific downstream tasks.
60
+
61
+ ## Limitations and Ethical Considerations
62
+
63
+ * **Unspecified Origin**: Without an official dataset card from the author, the precise origin and collection methodology of the sequences are unknown. This may impact reproducibility and potential biases.
64
+ * **Licensing**: The license is currently unspecified. For any public or commercial use, it is necessary to verify the terms with the author, David Cechak, on Hugging Face.
65
+
66
+ ## How to Get the Dataset
67
+
68
+ You can easily load this dataset from the Hugging Face Hub using the `datasets` library:
69
+
70
+ ```python
71
+ from datasets import load_dataset
72
+
73
+ # Load the tokenized dataset
74
+ dataset = load_dataset("davidcechak/Mouse_DNA_v0_DNABert6tokenized")
75
+
76
+ # Access the training split
77
+ train_dataset = dataset["train"]