File size: 12,556 Bytes
76be2de
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
# preprocess_data.py
import pandas as pd
import numpy as np
import re
import nltk
from nltk.corpus import stopwords
import string # To remove punctuation more easily
import os # To check for file existence

# --- Configuration ---
# Adjust these paths if your files are located elsewhere
TRAIN_TSV_PATH = 'train.tsv'
TEST_TSV_PATH = 'test.tsv'
FAKE_CSV_PATH = 'Fake.csv' # Adjust filename if needed (e.g., fake.csv)
TRUE_CSV_PATH = 'True.csv' # Adjust filename if needed (e.g., true.csv)
OUTPUT_CSV_PATH = 'combined_preprocessed_news.csv'

print("--- Data Preprocessing Script Started ---")

# --- Check if input files exist ---
required_files = [TRAIN_TSV_PATH, TEST_TSV_PATH, FAKE_CSV_PATH, TRUE_CSV_PATH]
files_exist = True
for f_path in required_files:
    if not os.path.exists(f_path):
        print(f"Error: Required input file not found: {f_path}")
        files_exist = False

if not files_exist:
    print("\nPlease ensure all required input files are in the same directory as the script.")
    print("Script aborted.")
    exit() # Stop the script if files are missing
else:
    print("All required input files found.")

# Define column names for the TSV files based on the description
tsv_column_names = [
    'id', 'label', 'statement', 'subject', 'speaker', 'speaker_job_title',
    'state_info', 'party_affiliation', 'barely_true_counts', 'false_counts',
    'half_true_counts', 'mostly_true_counts', 'pants_on_fire_counts', 'context'
]

# Define the target columns for the final unified dataset
final_columns = [
    'text', 'label', 'subject', 'speaker', 'speaker_job_title',
    'state_info', 'party_affiliation', 'context', 'date', 'source',
    'processed_text', 'encoded_label'
    # Add credit history counts if desired
    # 'barely_true_counts', 'false_counts', 'half_true_counts',
    # 'mostly_true_counts', 'pants_on_fire_counts'
]


# --- Download NLTK data (if not already present) ---
try:
    nltk.data.find('corpora/stopwords')
    print("\nNLTK stopwords found.")
except nltk.downloader.DownloadError:
    print("\nNLTK stopwords not found. Downloading...")
    try:
        nltk.download('stopwords')
        print("NLTK stopwords downloaded.")
    except Exception as e:
        print(f"Error downloading NLTK stopwords: {e}")
        print("Please check your internet connection or firewall settings.")
        print("Script aborted.")
        exit() # Stop if download fails

stop_words = set(stopwords.words('english'))

# --- Text Preprocessing Function ---
def preprocess_text(text):
    """Cleans and preprocesses text data."""
    if not isinstance(text, str):
        return "" # Return empty string for non-string inputs (like NaN)

    # 1. Convert to lowercase
    text = text.lower()

    # 2. Remove URLs
    text = re.sub(r'http\S+|www\S+|https\S+', '', text, flags=re.MULTILINE)

    # 3. Remove HTML tags (basic)
    text = re.sub(r'<.*?>', '', text)

    # 4. Remove text in square brackets (often metadata like [Reuters])
    text = re.sub(r'\[.*?\]', '', text)

    # 5. Remove punctuation more comprehensively
    text = text.translate(str.maketrans('', '', string.punctuation))

    # 6. Remove numbers (optional - consider if numbers are important)
    # text = re.sub(r'\d+', '', text)

    # 7. Remove extra whitespace
    text = re.sub(r'\s+', ' ', text).strip()

    # 8. Remove stopwords (optional - uncomment if desired)
    # tokens = text.split()
    # tokens = [word for word in tokens if word not in stop_words]
    # text = ' '.join(tokens)

    # 9. Lemmatization/Stemming (Optional - would require more NLTK setup)
    # ...

    return text

# --- Load and Process LIAR Dataset (TSV files) ---
print("\nProcessing LIAR dataset (train.tsv, test.tsv)...")
try:
    df_train_tsv = pd.read_csv(TRAIN_TSV_PATH, sep='\t', header=None, names=tsv_column_names)
    df_test_tsv = pd.read_csv(TEST_TSV_PATH, sep='\t', header=None, names=tsv_column_names)
    df_liar = pd.concat([df_train_tsv, df_test_tsv], ignore_index=True)
    print(f"LIAR dataset loaded. Shape: {df_liar.shape}")

    # Map LIAR labels to binary 'REAL'/'FAKE'
    # Categories considered REAL: true, mostly-true, half-true
    # Categories considered FAKE: false, barely-true, pants-fire
    label_mapping_liar = {
        'true': 'REAL',
        'mostly-true': 'REAL',
        'half-true': 'REAL',
        'false': 'FAKE',
        'barely-true': 'FAKE',
        'pants-fire': 'FAKE'
    }
    df_liar['label'] = df_liar['label'].map(label_mapping_liar)

    # Rename 'statement' to 'text' for consistency
    df_liar.rename(columns={'statement': 'text'}, inplace=True)

    # Add source column
    df_liar['source'] = 'LIAR'

    # Select relevant columns (add credit counts here if needed)
    liar_cols_to_keep = [
        'text', 'label', 'subject', 'speaker', 'speaker_job_title',
        'state_info', 'party_affiliation', 'context', 'source'
        # Add credit counts here if needed
        # 'barely_true_counts', 'false_counts', 'half_true_counts',
        # 'mostly_true_counts', 'pants_on_fire_counts'
    ]
    # Ensure only existing columns are selected
    liar_cols_to_keep = [col for col in liar_cols_to_keep if col in df_liar.columns]
    df_liar = df_liar[liar_cols_to_keep]
    print(f"LIAR dataset processed. Shape: {df_liar.shape}")

except FileNotFoundError as e:
    print(f"Error loading TSV file: {e}. Please check file paths.")
    df_liar = pd.DataFrame() # Create empty df to avoid error later
except Exception as e:
    print(f"An error occurred processing TSV files: {e}")
    df_liar = pd.DataFrame()

# --- Load and Process Kaggle Fake/True News Dataset (CSV files) ---
print("\nProcessing Kaggle dataset (Fake.csv, True.csv)...")
dfs_kaggle = []
try:
    # Process Fake News
    df_fake_csv = pd.read_csv(FAKE_CSV_PATH)
    df_fake_csv['label'] = 'FAKE'
    df_fake_csv['source'] = 'KaggleFake'
    # Combine title and text, handling potential NaNs
    df_fake_csv['text'] = df_fake_csv['title'].fillna('') + " " + df_fake_csv['text'].fillna('')
    dfs_kaggle.append(df_fake_csv[['text', 'label', 'subject', 'date', 'source']])
    print(f"Loaded Fake.csv. Shape: {df_fake_csv.shape}")

    # Process True News
    df_true_csv = pd.read_csv(TRUE_CSV_PATH)
    df_true_csv['label'] = 'REAL'
    df_true_csv['source'] = 'KaggleTrue'
    # Combine title and text, handling potential NaNs
    df_true_csv['text'] = df_true_csv['title'].fillna('') + " " + df_true_csv['text'].fillna('')
    dfs_kaggle.append(df_true_csv[['text', 'label', 'subject', 'date', 'source']])
    print(f"Loaded True.csv. Shape: {df_true_csv.shape}")

    df_kaggle = pd.concat(dfs_kaggle, ignore_index=True)
    print(f"Kaggle dataset processed. Shape: {df_kaggle.shape}")

except FileNotFoundError as e:
    print(f"Error loading CSV file: {e}. Please check file paths.")
    df_kaggle = pd.DataFrame() # Create empty df
except Exception as e:
    print(f"An error occurred processing CSV files: {e}")
    df_kaggle = pd.DataFrame()

# --- Combine Datasets ---
print("\nCombining LIAR and Kaggle datasets...")
if not df_liar.empty or not df_kaggle.empty:
    df_combined = pd.concat([df_liar, df_kaggle], ignore_index=True, sort=False)
    print(f"Initial combined shape: {df_combined.shape}")

    # Handle potential duplicate columns if credit counts were kept (unlikely here)
    # df_combined = df_combined.loc[:,~df_combined.columns.duplicated()]

    # Fill missing values created by combining different structures
    # Fill text NAs first (critical)
    initial_text_na = df_combined['text'].isna().sum()
    if initial_text_na > 0:
        print(f"Warning: Found {initial_text_na} missing values in 'text' column. Dropping these rows.")
        df_combined.dropna(subset=['text'], inplace=True)

    # Fill missing labels (less likely after processing, but check)
    initial_label_na = df_combined['label'].isna().sum()
    if initial_label_na > 0:
          print(f"Warning: Found {initial_label_na} missing values in 'label' column. Dropping these rows.")
          df_combined.dropna(subset=['label'], inplace=True)


    # Fill other missing metadata with 'Unknown' or appropriate placeholders
    fill_values = {
        'subject': 'Unknown',
        'speaker': 'Unknown',
        'speaker_job_title': 'Unknown',
        'state_info': 'Unknown',
        'party_affiliation': 'Unknown',
        'context': 'Unknown',
        'date': 'Unknown' # Or pd.NaT if you want a proper date null
        # Add fill values for credit counts if they were kept
    }
    # Apply fillna only for columns that actually exist in the combined df
    for col, value in fill_values.items():
        if col in df_combined.columns:
            df_combined[col].fillna(value, inplace=True)


    print(f"Shape after initial combining & handling critical NAs: {df_combined.shape}")
    # print(f"Combined dataset columns: {df_combined.columns.tolist()}") # Uncomment to debug columns if needed
    # print(f"\nMissing values per column after fill:\n{df_combined.isnull().sum()}") # Uncomment to debug NAs

    # --- Apply Text Preprocessing ---
    print("\nApplying text preprocessing to 'text' column...")
    # Ensure the text column is string type before applying
    df_combined['text'] = df_combined['text'].astype(str)
    df_combined['processed_text'] = df_combined['text'].apply(preprocess_text)
    print("Text preprocessing complete.")

    # --- Encode Labels ---
    print("Encoding labels ('FAKE': 0, 'REAL': 1)...") # IMPORTANT MAPPING
    label_encoding_map = {'FAKE': 0, 'REAL': 1}
    df_combined['encoded_label'] = df_combined['label'].map(label_encoding_map)

    # Verify encoding worked
    label_encoding_na = df_combined['encoded_label'].isna().sum()
    invalid_labels = df_combined[df_combined['encoded_label'].isna()]['label'].unique()
    if label_encoding_na > 0:
          print(f"Warning: {label_encoding_na} rows could not be label encoded.")
          print(f"Invalid or unexpected labels found: {invalid_labels}. Dropping these rows.")
          # Optionally drop these rows or investigate further
          df_combined.dropna(subset=['encoded_label'], inplace=True)
          df_combined['encoded_label'] = df_combined['encoded_label'].astype(int) # Ensure int type after dropna
    else:
      df_combined['encoded_label'] = df_combined['encoded_label'].astype(int)


    print("Label encoding complete.")
    print(f"\nValue counts for 'label':\n{df_combined['label'].value_counts()}")
    print(f"\nValue counts for 'encoded_label':\n{df_combined['encoded_label'].value_counts()}")


    # --- Final Touches ---
    # Drop rows where processed_text is empty after cleaning
    empty_processed = (df_combined['processed_text'] == '').sum()
    if empty_processed > 0:
        print(f"Dropping {empty_processed} rows where processed_text became empty after cleaning.")
        df_combined = df_combined[df_combined['processed_text'] != '']

    # Select and reorder columns for the final output
    # Ensure all desired final columns exist before selecting
    final_columns_present = [col for col in final_columns if col in df_combined.columns]
    df_final = df_combined[final_columns_present].copy()

    # Drop duplicates based on the processed text to avoid identical articles
    initial_rows = len(df_final)
    df_final.drop_duplicates(subset=['processed_text'], keep='first', inplace=True)
    rows_dropped = initial_rows - len(df_final)
    if rows_dropped > 0:
        print(f"Dropped {rows_dropped} duplicate rows based on 'processed_text'.")


    print(f"\nFinal dataset shape: {df_final.shape}")
    print(f"Final dataset columns: {df_final.columns.tolist()}")
    # print("\nFinal dataset info:") # Uncomment for more detail
    # df_final.info()
    # print("\nFinal dataset head:") # Uncomment for more detail
    # print(df_final.head())

    # --- Save the Preprocessed Data ---
    try:
        df_final.to_csv(OUTPUT_CSV_PATH, index=False)
        print(f"\nSuccessfully saved combined and preprocessed data to '{OUTPUT_CSV_PATH}'")
    except Exception as e:
        print(f"\nError saving the final CSV file: {e}")

else:
    print("\nCould not process input files. Combined dataset is empty. Nothing to save.")

print("\n--- Data Preprocessing Script Finished ---")