File size: 4,420 Bytes
e035723
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4ea4cd7
 
 
 
 
e035723
 
 
 
 
 
4ea4cd7
e035723
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
import os
from os.path import exists

import datasets
from datasets.tasks import TextClassification
from datasets import load_dataset

import numpy as np 
import json

from transformers import AutoTokenizer


logger = datasets.logging.get_logger(__name__)

## Constants
USE_FULL_DATASET = True
PROJECT_PATH = "./"




class Sentiment(datasets.GeneratorBasedBuilder):
    '''Custom Dataset created using the HuggingFace api so we can 
    use all of their's api on the dataset'''
    def _info(self):
        class_names = ["negative", "positive"]
        return datasets.DatasetInfo(
            description="Our nice dataset in HF format",
            features=datasets.Features(
                {"text": datasets.Value("string"), 
                "labels": datasets.ClassLabel(num_classes=2, names=class_names)}  #  Value("int32")
            ),
            supervised_keys=("text", "labels"),
        )
    
    def _define_columns(self,example):
        text_splited = example["text"].split('\t')
        return {"text": text_splited[1].strip(), "labels": int(text_splited[0])}
    
    def _split_generators(self, _):
        """Returns SplitGenerators."""

        data_dir = "./"

        data = load_dataset("text", data_files="./HF_data.txt")
        data = data.map(self._define_columns)

        texts_dataset_clean = data["train"].train_test_split(train_size=0.95, seed=12345)
        # Rename the default "test" split to "validation"
        texts_dataset_clean["validation"] = texts_dataset_clean.pop("test")


        for split, dataset in texts_dataset_clean.items():
            dataset.to_json(data_dir + f"twitter-sentiment-analysis-{split}.jsonl")


        return [
            datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"filepath": os.path.join(data_dir, "twitter-sentiment-analysis-train.jsonl")}),
            datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={"filepath": os.path.join(data_dir, "twitter-sentiment-analysis-validation.jsonl")}),
        ]   

    def _generate_examples(self, filepath):
        """This function returns the examples in the raw (text) form."""
        logger.info("generating examples from = %s", filepath)
        with open(filepath, encoding="utf-8") as f:
            for key, row in enumerate(f):
                data = json.loads(row)
                yield key, {
                    "text": data["text"],
                    "labels": data["labels"],
                }


def read_file(file_name_label_tuple):
    fname, label = file_name_label_tuple
    tweets, labels = [], []
    with open(fname, 'r', encoding='utf-8') as f:
        tweets = f.readlines()
    
    labels = [label] * (len(tweets))

    return(tweets, labels)


def load_train_data():

    if USE_FULL_DATASET == True:
        X_train_neg_path = PROJECT_PATH + "train_neg_full.txt"
        X_train_pos_path = PROJECT_PATH + "train_pos_full.txt"
        
    else:
        X_train_neg_path = PROJECT_PATH + "train_neg.txt"
        X_train_pos_path = PROJECT_PATH + "train_pos.txt"
    
    tweets, labels = read_file((X_train_neg_path, 0))
    tweets = list(set(tweets))
    labels = labels[:len(tweets)]
    print("There are ", len(tweets), " negative tweets after removing the duplicates.")
    
    tweets_2, labels_2 = read_file((X_train_pos_path, 1))
    tweets_2 = list(set(tweets_2))
    labels_2 = labels_2[:len(tweets_2)]
    print("There are ", len(tweets_2), " positive tweets after removing the duplicates.")

    
    tweets += tweets_2
    tweets_2 = []
    del(tweets_2)
    labels += labels_2
    labels_2 = []
    del(labels_2)
    print(f"Loaded {len(tweets)} tweets!")

    tweets, labels = np.array(tweets), np.array(labels)
    print(tweets)

    # To shuffle the data before cerating the .txt file dataset
    nb_of_samples = len(tweets)
    shuffled_indices = np.random.permutation(nb_of_samples)
    tweets = tweets[shuffled_indices]
    labels = labels[shuffled_indices]

    print("Number of indices for training: ", len(shuffled_indices))

    return tweets, labels


def create_data_file(tweets, labels):
    
    with open("HF_data.txt", "wb") as f:
        for i in range(len(tweets)):
            # print(tweets[i])
            f.write(f"{labels[i]} \t {tweets[i]}".encode('utf-8'))

    
    

def main():
    tweets, labels = load_train_data()
    create_data_file(tweets, labels)

if __name__ == "__main__":
    main()