eoinf commited on
Commit
60290e4
·
verified ·
1 Parent(s): 33b409f

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +60 -15
README.md CHANGED
@@ -1,17 +1,62 @@
1
  ---
2
- dataset_info:
3
- features:
4
- - name: tokens
5
- list: int64
6
- splits:
7
- - name: train
8
- num_bytes: 80066724
9
- num_examples: 9769
10
- download_size: 17920074
11
- dataset_size: 80066724
12
- configs:
13
- - config_name: default
14
- data_files:
15
- - split: train
16
- path: data/train-*
17
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: mit
3
+ task_categories:
4
+ - text-generation
5
+ language:
6
+ - en
7
+ tags:
8
+ - tokenized
9
+ - language-modeling
10
+ size_categories:
11
+ - 1K<n<10K
 
 
 
 
 
12
  ---
13
+ # Dataset Card for eoinf/tokenized_dataset_test8
14
+
15
+ ## Original dataset
16
+ Original dataset: Salesforce/wikitext
17
+
18
+ ## Dataset Details
19
+
20
+ - **Total Tokens**: 10,003,456
21
+ - **Total Sequences**: 9,769
22
+ - **Context Length**: 1024 tokens
23
+ - **Tokenizer**: meta-llama/Llama-2-7b-hf
24
+ - **Format**: Each example contains a single field `tokens` with a list of 1024 token IDs
25
+
26
+ ## Preprocessing
27
+
28
+ Each document was:
29
+ 1. Tokenized using the meta-llama/Llama-2-7b-hf tokenizer
30
+ 2. Prefixed with a BOS (beginning of sequence) token
31
+ 3. Suffixed with an EOS (end of sequence) token
32
+ 4. Packed into fixed-length sequences of 1024 tokens
33
+
34
+ ## Usage
35
+ ```python
36
+ from datasets import load_dataset
37
+
38
+ # Load the dataset
39
+ dataset = load_dataset("eoinf/tokenized_dataset_test8")
40
+
41
+ # Access training data
42
+ train_data = dataset["train"]
43
+ print(train_data[0]["tokens"]) # First sequence
44
+ ```
45
+
46
+ ## Use with PyTorch
47
+ ```python
48
+ import torch
49
+ from datasets import load_dataset
50
+ from torch.utils.data import DataLoader
51
+
52
+ dataset = load_dataset("eoinf/tokenized_dataset_test8", split="train")
53
+
54
+ # Convert to PyTorch tensors
55
+ dataset.set_format(type="torch", columns=["tokens"])
56
+
57
+ # Create DataLoader
58
+ dataloader = DataLoader(dataset, batch_size=32, shuffle=True)
59
+
60
+ for batch in dataloader:
61
+ tokens = batch["tokens"]
62
+ ```