text
stringlengths 4
39
|
|---|
1 This
|
2 engaging
|
3 collectible
|
4 miniature
|
5 hardcover
|
6 of
|
7 the
|
8 Orson
|
9 Scott
|
10 Card
|
11 classic
|
12 and
|
13 worldwide
|
14 bestselling
|
15 novel
|
16 Enders
|
17 Game
|
18 makes
|
19 an
|
20 excellent
|
21 gift
|
22 for
|
23 anyones
|
24 science
|
25 fiction
|
26 library
|
27 Enders
|
28 Game
|
29 is
|
30 an
|
31 affecting
|
32 novelNew
|
33 York
|
34 Times
|
35 Book
|
36 Review
|
37 Once
|
38 again
|
39 Earth
|
40 is
|
41 under
|
42 attack
|
43 An
|
44 alien
|
45 species
|
46 is
|
47 poised
|
48 for
|
49 a
|
50 final
|
51 assault
|
52 The
|
53 survival
|
54 of
|
55 humanity
|
56 depends
|
57 on
|
58 a
|
59 military
|
60 genius
|
61 who
|
62 can
|
63 defeat
|
64 the
|
65 aliens
|
66 But
|
67 who
|
68 Ender
|
69 Wiggin
|
70 Brilliant
|
71 Ruthless
|
72 Cunning
|
73 A
|
74 tactical
|
75 and
|
76 strategic
|
77 master
|
78 And
|
79 a
|
80 child
|
81 Recruited
|
82 for
|
83 military
|
84 training
|
85 by
|
86 the
|
87 world
|
88 government
|
89 Enders
|
90 childhood
|
91 ends
|
92 the
|
93 moment
|
94 he
|
95 enters
|
96 his
|
97 new
|
98 home
|
99 Battle
|
100 School
|
Dataset Overview
Dataset Name: Whitzz/EnglishDatasets
License: MIT
Description:
This dataset consists of 100,000+ English words consisting of many version ranging from small to huge, scraped from multiple sources. It can be used for fine-tuning language models, performing text processing tasks, or for applications like spell-checking, word categorization, and more.
How to Use
To use this dataset in Google Colab or any Python environment, follow these steps:
Step 1: Install the Required Library
The dataset is available through the datasets library by Hugging Face. First, you need to install the library by running the following command:
!pip install datasets
Step 2: Load the Dataset
Once the library is installed, you can proceed to load the "Whitzz/EnglishDatasets" dataset. Here's how you can do it:
from datasets import load_dataset
# Load the Whitzz/EnglishDatasets
dataset = load_dataset('Whitzz/EnglishDatasets')
# Print the dataset to check its structure
print(dataset)
When you run this code, it will load the dataset and print a summary of its structure. The dataset might contain multiple splits, such as "train", "test", and "validation".
Step 3: Print Dataset Entries
To view the actual data, you can print a small portion of the dataset. For instance, you can print the first five entries from the "train" split like this:
# Print the first 5 entries in the 'train' split
print(dataset['train'][:5])
Result: DatasetDict({
train: Dataset({
features: ['text'],
num_rows: 100000
})
})
Additional Operations You Can Perform:
Here are some other functions you might find useful for exploring or processing the dataset:
1. Filter the Dataset:
You can also filter the dataset based on certain conditions. For example, filtering words that start with a specific letter:
# Filter words starting with 'a'
filtered_data = dataset['train'].filter(lambda example: example['text'].startswith('a'))
# Print the first 5 filtered examples
print(filtered_data[:5])
Use the Dataset for Training:
Once you load and explore the dataset, you can use it to fine-tune language models (e.g., GPT-3, BERT). For instance, you can prepare the data by tokenizing it and then feeding it into a model.
Example (using transformers library for tokenization):
from transformers import AutoTokenizer
# Load a pre-trained tokenizer (e.g., BERT tokenizer)
tokenizer = AutoTokenizer.from_pretrained('bert-base-uncased')
# Tokenize the dataset (just an example)
def tokenize_function(examples):
return tokenizer(examples['text'], padding='max_length', truncation=True)
tokenized_datasets = dataset.map(tokenize_function, batched=True)
# Now tokenized_datasets is ready for fine-tuning or further processing
- Downloads last month
- 5