File size: 12,556 Bytes
76be2de |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 |
# preprocess_data.py
import pandas as pd
import numpy as np
import re
import nltk
from nltk.corpus import stopwords
import string # To remove punctuation more easily
import os # To check for file existence
# --- Configuration ---
# Adjust these paths if your files are located elsewhere
TRAIN_TSV_PATH = 'train.tsv'
TEST_TSV_PATH = 'test.tsv'
FAKE_CSV_PATH = 'Fake.csv' # Adjust filename if needed (e.g., fake.csv)
TRUE_CSV_PATH = 'True.csv' # Adjust filename if needed (e.g., true.csv)
OUTPUT_CSV_PATH = 'combined_preprocessed_news.csv'
print("--- Data Preprocessing Script Started ---")
# --- Check if input files exist ---
required_files = [TRAIN_TSV_PATH, TEST_TSV_PATH, FAKE_CSV_PATH, TRUE_CSV_PATH]
files_exist = True
for f_path in required_files:
if not os.path.exists(f_path):
print(f"Error: Required input file not found: {f_path}")
files_exist = False
if not files_exist:
print("\nPlease ensure all required input files are in the same directory as the script.")
print("Script aborted.")
exit() # Stop the script if files are missing
else:
print("All required input files found.")
# Define column names for the TSV files based on the description
tsv_column_names = [
'id', 'label', 'statement', 'subject', 'speaker', 'speaker_job_title',
'state_info', 'party_affiliation', 'barely_true_counts', 'false_counts',
'half_true_counts', 'mostly_true_counts', 'pants_on_fire_counts', 'context'
]
# Define the target columns for the final unified dataset
final_columns = [
'text', 'label', 'subject', 'speaker', 'speaker_job_title',
'state_info', 'party_affiliation', 'context', 'date', 'source',
'processed_text', 'encoded_label'
# Add credit history counts if desired
# 'barely_true_counts', 'false_counts', 'half_true_counts',
# 'mostly_true_counts', 'pants_on_fire_counts'
]
# --- Download NLTK data (if not already present) ---
try:
nltk.data.find('corpora/stopwords')
print("\nNLTK stopwords found.")
except nltk.downloader.DownloadError:
print("\nNLTK stopwords not found. Downloading...")
try:
nltk.download('stopwords')
print("NLTK stopwords downloaded.")
except Exception as e:
print(f"Error downloading NLTK stopwords: {e}")
print("Please check your internet connection or firewall settings.")
print("Script aborted.")
exit() # Stop if download fails
stop_words = set(stopwords.words('english'))
# --- Text Preprocessing Function ---
def preprocess_text(text):
"""Cleans and preprocesses text data."""
if not isinstance(text, str):
return "" # Return empty string for non-string inputs (like NaN)
# 1. Convert to lowercase
text = text.lower()
# 2. Remove URLs
text = re.sub(r'http\S+|www\S+|https\S+', '', text, flags=re.MULTILINE)
# 3. Remove HTML tags (basic)
text = re.sub(r'<.*?>', '', text)
# 4. Remove text in square brackets (often metadata like [Reuters])
text = re.sub(r'\[.*?\]', '', text)
# 5. Remove punctuation more comprehensively
text = text.translate(str.maketrans('', '', string.punctuation))
# 6. Remove numbers (optional - consider if numbers are important)
# text = re.sub(r'\d+', '', text)
# 7. Remove extra whitespace
text = re.sub(r'\s+', ' ', text).strip()
# 8. Remove stopwords (optional - uncomment if desired)
# tokens = text.split()
# tokens = [word for word in tokens if word not in stop_words]
# text = ' '.join(tokens)
# 9. Lemmatization/Stemming (Optional - would require more NLTK setup)
# ...
return text
# --- Load and Process LIAR Dataset (TSV files) ---
print("\nProcessing LIAR dataset (train.tsv, test.tsv)...")
try:
df_train_tsv = pd.read_csv(TRAIN_TSV_PATH, sep='\t', header=None, names=tsv_column_names)
df_test_tsv = pd.read_csv(TEST_TSV_PATH, sep='\t', header=None, names=tsv_column_names)
df_liar = pd.concat([df_train_tsv, df_test_tsv], ignore_index=True)
print(f"LIAR dataset loaded. Shape: {df_liar.shape}")
# Map LIAR labels to binary 'REAL'/'FAKE'
# Categories considered REAL: true, mostly-true, half-true
# Categories considered FAKE: false, barely-true, pants-fire
label_mapping_liar = {
'true': 'REAL',
'mostly-true': 'REAL',
'half-true': 'REAL',
'false': 'FAKE',
'barely-true': 'FAKE',
'pants-fire': 'FAKE'
}
df_liar['label'] = df_liar['label'].map(label_mapping_liar)
# Rename 'statement' to 'text' for consistency
df_liar.rename(columns={'statement': 'text'}, inplace=True)
# Add source column
df_liar['source'] = 'LIAR'
# Select relevant columns (add credit counts here if needed)
liar_cols_to_keep = [
'text', 'label', 'subject', 'speaker', 'speaker_job_title',
'state_info', 'party_affiliation', 'context', 'source'
# Add credit counts here if needed
# 'barely_true_counts', 'false_counts', 'half_true_counts',
# 'mostly_true_counts', 'pants_on_fire_counts'
]
# Ensure only existing columns are selected
liar_cols_to_keep = [col for col in liar_cols_to_keep if col in df_liar.columns]
df_liar = df_liar[liar_cols_to_keep]
print(f"LIAR dataset processed. Shape: {df_liar.shape}")
except FileNotFoundError as e:
print(f"Error loading TSV file: {e}. Please check file paths.")
df_liar = pd.DataFrame() # Create empty df to avoid error later
except Exception as e:
print(f"An error occurred processing TSV files: {e}")
df_liar = pd.DataFrame()
# --- Load and Process Kaggle Fake/True News Dataset (CSV files) ---
print("\nProcessing Kaggle dataset (Fake.csv, True.csv)...")
dfs_kaggle = []
try:
# Process Fake News
df_fake_csv = pd.read_csv(FAKE_CSV_PATH)
df_fake_csv['label'] = 'FAKE'
df_fake_csv['source'] = 'KaggleFake'
# Combine title and text, handling potential NaNs
df_fake_csv['text'] = df_fake_csv['title'].fillna('') + " " + df_fake_csv['text'].fillna('')
dfs_kaggle.append(df_fake_csv[['text', 'label', 'subject', 'date', 'source']])
print(f"Loaded Fake.csv. Shape: {df_fake_csv.shape}")
# Process True News
df_true_csv = pd.read_csv(TRUE_CSV_PATH)
df_true_csv['label'] = 'REAL'
df_true_csv['source'] = 'KaggleTrue'
# Combine title and text, handling potential NaNs
df_true_csv['text'] = df_true_csv['title'].fillna('') + " " + df_true_csv['text'].fillna('')
dfs_kaggle.append(df_true_csv[['text', 'label', 'subject', 'date', 'source']])
print(f"Loaded True.csv. Shape: {df_true_csv.shape}")
df_kaggle = pd.concat(dfs_kaggle, ignore_index=True)
print(f"Kaggle dataset processed. Shape: {df_kaggle.shape}")
except FileNotFoundError as e:
print(f"Error loading CSV file: {e}. Please check file paths.")
df_kaggle = pd.DataFrame() # Create empty df
except Exception as e:
print(f"An error occurred processing CSV files: {e}")
df_kaggle = pd.DataFrame()
# --- Combine Datasets ---
print("\nCombining LIAR and Kaggle datasets...")
if not df_liar.empty or not df_kaggle.empty:
df_combined = pd.concat([df_liar, df_kaggle], ignore_index=True, sort=False)
print(f"Initial combined shape: {df_combined.shape}")
# Handle potential duplicate columns if credit counts were kept (unlikely here)
# df_combined = df_combined.loc[:,~df_combined.columns.duplicated()]
# Fill missing values created by combining different structures
# Fill text NAs first (critical)
initial_text_na = df_combined['text'].isna().sum()
if initial_text_na > 0:
print(f"Warning: Found {initial_text_na} missing values in 'text' column. Dropping these rows.")
df_combined.dropna(subset=['text'], inplace=True)
# Fill missing labels (less likely after processing, but check)
initial_label_na = df_combined['label'].isna().sum()
if initial_label_na > 0:
print(f"Warning: Found {initial_label_na} missing values in 'label' column. Dropping these rows.")
df_combined.dropna(subset=['label'], inplace=True)
# Fill other missing metadata with 'Unknown' or appropriate placeholders
fill_values = {
'subject': 'Unknown',
'speaker': 'Unknown',
'speaker_job_title': 'Unknown',
'state_info': 'Unknown',
'party_affiliation': 'Unknown',
'context': 'Unknown',
'date': 'Unknown' # Or pd.NaT if you want a proper date null
# Add fill values for credit counts if they were kept
}
# Apply fillna only for columns that actually exist in the combined df
for col, value in fill_values.items():
if col in df_combined.columns:
df_combined[col].fillna(value, inplace=True)
print(f"Shape after initial combining & handling critical NAs: {df_combined.shape}")
# print(f"Combined dataset columns: {df_combined.columns.tolist()}") # Uncomment to debug columns if needed
# print(f"\nMissing values per column after fill:\n{df_combined.isnull().sum()}") # Uncomment to debug NAs
# --- Apply Text Preprocessing ---
print("\nApplying text preprocessing to 'text' column...")
# Ensure the text column is string type before applying
df_combined['text'] = df_combined['text'].astype(str)
df_combined['processed_text'] = df_combined['text'].apply(preprocess_text)
print("Text preprocessing complete.")
# --- Encode Labels ---
print("Encoding labels ('FAKE': 0, 'REAL': 1)...") # IMPORTANT MAPPING
label_encoding_map = {'FAKE': 0, 'REAL': 1}
df_combined['encoded_label'] = df_combined['label'].map(label_encoding_map)
# Verify encoding worked
label_encoding_na = df_combined['encoded_label'].isna().sum()
invalid_labels = df_combined[df_combined['encoded_label'].isna()]['label'].unique()
if label_encoding_na > 0:
print(f"Warning: {label_encoding_na} rows could not be label encoded.")
print(f"Invalid or unexpected labels found: {invalid_labels}. Dropping these rows.")
# Optionally drop these rows or investigate further
df_combined.dropna(subset=['encoded_label'], inplace=True)
df_combined['encoded_label'] = df_combined['encoded_label'].astype(int) # Ensure int type after dropna
else:
df_combined['encoded_label'] = df_combined['encoded_label'].astype(int)
print("Label encoding complete.")
print(f"\nValue counts for 'label':\n{df_combined['label'].value_counts()}")
print(f"\nValue counts for 'encoded_label':\n{df_combined['encoded_label'].value_counts()}")
# --- Final Touches ---
# Drop rows where processed_text is empty after cleaning
empty_processed = (df_combined['processed_text'] == '').sum()
if empty_processed > 0:
print(f"Dropping {empty_processed} rows where processed_text became empty after cleaning.")
df_combined = df_combined[df_combined['processed_text'] != '']
# Select and reorder columns for the final output
# Ensure all desired final columns exist before selecting
final_columns_present = [col for col in final_columns if col in df_combined.columns]
df_final = df_combined[final_columns_present].copy()
# Drop duplicates based on the processed text to avoid identical articles
initial_rows = len(df_final)
df_final.drop_duplicates(subset=['processed_text'], keep='first', inplace=True)
rows_dropped = initial_rows - len(df_final)
if rows_dropped > 0:
print(f"Dropped {rows_dropped} duplicate rows based on 'processed_text'.")
print(f"\nFinal dataset shape: {df_final.shape}")
print(f"Final dataset columns: {df_final.columns.tolist()}")
# print("\nFinal dataset info:") # Uncomment for more detail
# df_final.info()
# print("\nFinal dataset head:") # Uncomment for more detail
# print(df_final.head())
# --- Save the Preprocessed Data ---
try:
df_final.to_csv(OUTPUT_CSV_PATH, index=False)
print(f"\nSuccessfully saved combined and preprocessed data to '{OUTPUT_CSV_PATH}'")
except Exception as e:
print(f"\nError saving the final CSV file: {e}")
else:
print("\nCould not process input files. Combined dataset is empty. Nothing to save.")
print("\n--- Data Preprocessing Script Finished ---") |