Dataset Viewer
Auto-converted to Parquet Duplicate
text
stringlengths
4
39
1 This
2 engaging
3 collectible
4 miniature
5 hardcover
6 of
7 the
8 Orson
9 Scott
10 Card
11 classic
12 and
13 worldwide
14 bestselling
15 novel
16 Enders
17 Game
18 makes
19 an
20 excellent
21 gift
22 for
23 anyones
24 science
25 fiction
26 library
27 Enders
28 Game
29 is
30 an
31 affecting
32 novelNew
33 York
34 Times
35 Book
36 Review
37 Once
38 again
39 Earth
40 is
41 under
42 attack
43 An
44 alien
45 species
46 is
47 poised
48 for
49 a
50 final
51 assault
52 The
53 survival
54 of
55 humanity
56 depends
57 on
58 a
59 military
60 genius
61 who
62 can
63 defeat
64 the
65 aliens
66 But
67 who
68 Ender
69 Wiggin
70 Brilliant
71 Ruthless
72 Cunning
73 A
74 tactical
75 and
76 strategic
77 master
78 And
79 a
80 child
81 Recruited
82 for
83 military
84 training
85 by
86 the
87 world
88 government
89 Enders
90 childhood
91 ends
92 the
93 moment
94 he
95 enters
96 his
97 new
98 home
99 Battle
100 School
End of preview. Expand in Data Studio
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/datasets-cards)

Dataset Overview

Dataset Name: Whitzz/EnglishDatasets
License: MIT
Description:
This dataset consists of 100,000+ English words consisting of many version ranging from small to huge, scraped from multiple sources. It can be used for fine-tuning language models, performing text processing tasks, or for applications like spell-checking, word categorization, and more.

How to Use

To use this dataset in Google Colab or any Python environment, follow these steps:

Step 1: Install the Required Library

The dataset is available through the datasets library by Hugging Face. First, you need to install the library by running the following command:

!pip install datasets

Step 2: Load the Dataset

Once the library is installed, you can proceed to load the "Whitzz/EnglishDatasets" dataset. Here's how you can do it:

from datasets import load_dataset

# Load the Whitzz/EnglishDatasets
dataset = load_dataset('Whitzz/EnglishDatasets')

# Print the dataset to check its structure
print(dataset)

When you run this code, it will load the dataset and print a summary of its structure. The dataset might contain multiple splits, such as "train", "test", and "validation".

Step 3: Print Dataset Entries

To view the actual data, you can print a small portion of the dataset. For instance, you can print the first five entries from the "train" split like this:

# Print the first 5 entries in the 'train' split
print(dataset['train'][:5])
Result: DatasetDict({
    train: Dataset({
        features: ['text'],
        num_rows: 100000
    })
})

Additional Operations You Can Perform:

Here are some other functions you might find useful for exploring or processing the dataset:

1. Filter the Dataset:

You can also filter the dataset based on certain conditions. For example, filtering words that start with a specific letter:

# Filter words starting with 'a'
filtered_data = dataset['train'].filter(lambda example: example['text'].startswith('a'))

# Print the first 5 filtered examples
print(filtered_data[:5])

Use the Dataset for Training:

Once you load and explore the dataset, you can use it to fine-tune language models (e.g., GPT-3, BERT). For instance, you can prepare the data by tokenizing it and then feeding it into a model.

Example (using transformers library for tokenization):

from transformers import AutoTokenizer

# Load a pre-trained tokenizer (e.g., BERT tokenizer)
tokenizer = AutoTokenizer.from_pretrained('bert-base-uncased')

# Tokenize the dataset (just an example)
def tokenize_function(examples):
    return tokenizer(examples['text'], padding='max_length', truncation=True)

tokenized_datasets = dataset.map(tokenize_function, batched=True)

# Now tokenized_datasets is ready for fine-tuning or further processing
Downloads last month
5