Spaces:
Running
Running
File size: 12,638 Bytes
0304d75 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 | """
tests/test_data_preprocessing.py
=================================
Unit tests for the data_preprocessing module.
Tests use synthetic in-memory DataFrames so no real dataset files are needed.
"""
import os
import tempfile
import pandas as pd
import pytest
# Import functions under test
from data_preprocessing import (
_CONTRAST_SAMPLES,
_find_column,
clean_text,
load_dreaddit,
load_happy_neutral,
load_reddit_combi,
load_reddit_title,
load_stressed_tweets,
load_twitter_full,
merge_datasets,
)
# ---------------------------------------------------------------------------
# Helpers
# ---------------------------------------------------------------------------
def _write_csv(df: pd.DataFrame, path: str) -> None:
os.makedirs(os.path.dirname(path), exist_ok=True)
df.to_csv(path, index=False)
# ---------------------------------------------------------------------------
# Tests for _find_column
# ---------------------------------------------------------------------------
class TestFindColumn:
def test_exact_match(self):
df = pd.DataFrame(columns=["text", "label", "extra"])
assert _find_column(df, ["text"]) == "text"
def test_case_insensitive_match(self):
df = pd.DataFrame(columns=["Text", "Label"])
assert _find_column(df, ["text"]) == "Text"
def test_first_candidate_wins(self):
df = pd.DataFrame(columns=["selftext", "text"])
assert _find_column(df, ["text", "selftext"]) == "text"
def test_fallback_candidate(self):
df = pd.DataFrame(columns=["selftext", "label"])
assert _find_column(df, ["text", "selftext"]) == "selftext"
def test_missing_raises(self):
df = pd.DataFrame(columns=["foo", "bar"])
with pytest.raises(KeyError):
_find_column(df, ["text", "label"])
# ---------------------------------------------------------------------------
# Tests for clean_text
# ---------------------------------------------------------------------------
class TestCleanText:
def test_strips_whitespace(self):
# clean_text now takes a string and returns a string
assert clean_text(" hello ").strip() == "hello"
assert clean_text("world ").strip() == "world"
def test_empty_string_becomes_empty(self):
# Empty / whitespace-only inputs return an empty string
assert clean_text("") == ""
assert clean_text(" ") == ""
def test_preserves_content_words(self):
# Content words survive stop-word filtering; "I" and "feel" are
# handled differently — "i" is a stop word, "feel" is kept.
result = clean_text("I feel stressed")
assert "stressed" in result
assert "feel" in result
def test_removes_stop_words(self):
# Pure stop words should be filtered
result = clean_text("the a an is are")
# After filtering only stop words, falls back to lowercased original
# (no content words to return); important thing: no error.
assert isinstance(result, str)
def test_preserves_contrast_words(self):
# 'not', 'but', 'however' must survive filtering
result = clean_text("I am not happy but okay")
assert "not" in result
assert "but" in result
# ---------------------------------------------------------------------------
# Tests for individual dataset loaders
# ---------------------------------------------------------------------------
class TestLoadDreaddit:
def test_loads_csv(self, tmp_path):
df = pd.DataFrame({"text": ["post1", "post2"], "label": [0, 1]})
csv_path = str(tmp_path / "dreaddit-train.csv")
df.to_csv(csv_path, index=False)
result = load_dreaddit(csv_path)
assert list(result.columns) == ["text", "label", "domain"]
assert len(result) == 2
assert all(result["domain"] == "reddit_long")
def test_loads_from_zip(self, tmp_path):
import zipfile
df = pd.DataFrame({"text": ["zipped post"], "label": [1]})
csv_name = "dreaddit-train.csv"
csv_in_zip = tmp_path / csv_name
df.to_csv(str(csv_in_zip), index=False)
zip_path = str(tmp_path / (csv_name + ".zip"))
with zipfile.ZipFile(zip_path, "w") as zf:
zf.write(str(csv_in_zip), csv_name)
# Remove the plain CSV so the loader falls back to ZIP
os.remove(str(csv_in_zip))
result = load_dreaddit(str(csv_in_zip))
assert len(result) == 1
assert result["label"].iloc[0] == 1
class TestLoadRedditCombi:
def test_basic_load(self, tmp_path):
df = pd.DataFrame({"text": ["combo post"], "label": [0]})
csv_path = str(tmp_path / "Reddit_Combi.csv")
df.to_csv(csv_path, index=False)
result = load_reddit_combi(csv_path)
assert len(result) == 1
assert result["domain"].iloc[0] == "reddit_long"
class TestLoadRedditTitle:
def test_basic_load(self, tmp_path):
df = pd.DataFrame({"title": ["My Title"], "label": [1]})
csv_path = str(tmp_path / "Reddit_Title.csv")
df.to_csv(csv_path, index=False)
result = load_reddit_title(csv_path)
assert len(result) == 1
assert result["text"].iloc[0] == "My Title"
assert result["domain"].iloc[0] == "reddit_short"
class TestLoadTwitterFull:
def test_basic_load(self, tmp_path):
df = pd.DataFrame({"text": ["tweet1", "tweet2"], "label": [0, 1]})
csv_path = str(tmp_path / "Twitter_Full.csv")
df.to_csv(csv_path, index=False)
result = load_twitter_full(csv_path)
assert len(result) == 2
assert all(result["domain"] == "twitter_short")
class TestLoadStressedTweets:
def test_implicit_label(self, tmp_path):
df = pd.DataFrame({"text": ["I am so stressed", "Everything is awful"]})
csv_path = str(tmp_path / "Stressed_Tweets.csv")
df.to_csv(csv_path, index=False)
result = load_stressed_tweets(csv_path)
assert len(result) == 2
assert all(result["label"] == 1)
assert all(result["domain"] == "twitter_short")
class TestLoadHappyNeutral:
def test_implicit_label(self, tmp_path):
df = pd.DataFrame({"text": ["I feel great", "What a wonderful day"]})
csv_path = str(tmp_path / "Happy_Neutral.csv")
df.to_csv(csv_path, index=False)
result = load_happy_neutral(csv_path)
assert len(result) == 2
assert all(result["label"] == 0)
assert all(result["domain"] == "happy_neutral")
# ---------------------------------------------------------------------------
# Tests for merge_datasets
# ---------------------------------------------------------------------------
class TestMergeDatasets:
def test_merges_multiple_datasets(self, tmp_path):
"""Merge 3 available datasets; 2 missing should be skipped.
The 12 hard-negative contrast samples are always appended."""
raw_dir = tmp_path / "raw"
raw_dir.mkdir()
out_path = str(tmp_path / "unified.csv")
# Create dreaddit
pd.DataFrame({"text": ["dreaddit post"], "label": [0]}).to_csv(
str(raw_dir / "dreaddit-train.csv"), index=False
)
# Create Twitter_Full
pd.DataFrame({"text": ["tweet text"], "label": [1]}).to_csv(
str(raw_dir / "Twitter_Full.csv"), index=False
)
# Create Stressed_Tweets
pd.DataFrame({"text": ["stressed tweet"]}).to_csv(
str(raw_dir / "Stressed_Tweets.csv"), index=False
)
result = merge_datasets(str(raw_dir), out_path)
n_contrast = len(_CONTRAST_SAMPLES)
assert len(result) == 3 + n_contrast
assert {"text", "label", "domain"}.issubset(result.columns)
assert os.path.isfile(out_path)
# Verify saved CSV matches
saved = pd.read_csv(out_path)
assert len(saved) == 3 + n_contrast
def test_drops_empty_text(self, tmp_path):
"""Rows with empty/whitespace text should be dropped.
Contrast samples are always appended after dropping."""
raw_dir = tmp_path / "raw"
raw_dir.mkdir()
out_path = str(tmp_path / "unified.csv")
pd.DataFrame(
{"text": ["valid", "", " ", "also valid"], "label": [0, 1, 0, 1]}
).to_csv(str(raw_dir / "dreaddit-train.csv"), index=False)
result = merge_datasets(str(raw_dir), out_path)
n_contrast = len(_CONTRAST_SAMPLES)
assert len(result) == 2 + n_contrast
# The dataset rows (before contrast samples) should contain the valid texts
dataset_texts = result["text"].tolist()
assert "valid" in dataset_texts
assert "also valid" in dataset_texts
def test_label_is_int(self, tmp_path):
"""Labels should always be int 0 or 1."""
raw_dir = tmp_path / "raw"
raw_dir.mkdir()
out_path = str(tmp_path / "unified.csv")
pd.DataFrame({"text": ["post"], "label": [1.0]}).to_csv(
str(raw_dir / "dreaddit-train.csv"), index=False
)
result = merge_datasets(str(raw_dir), out_path)
assert pd.api.types.is_integer_dtype(result["label"])
def test_no_datasets_exits(self, tmp_path):
"""If no datasets found, the function should exit."""
raw_dir = tmp_path / "empty"
raw_dir.mkdir()
out_path = str(tmp_path / "unified.csv")
with pytest.raises(SystemExit):
merge_datasets(str(raw_dir), out_path)
# ---------------------------------------------------------------------------
# Regression tests for _CONTRAST_SAMPLES
# ---------------------------------------------------------------------------
class TestContrastSamplesRegression:
"""Regression tests to catch mislabelled entries in _CONTRAST_SAMPLES.
These guard against future regressions where clearly stressed text
is accidentally labelled 0 or clearly non-stressed text is labelled 1.
"""
def test_deadline_sleep_deprivation_is_stressed(self):
"""'Deadlines piling up + not slept' must be label=1 (stressed)."""
texts_that_must_be_stressed = [
"deadlines are piling up and I cannot keep up",
"deadlines are piling up I have not slept properly in days",
"I have not slept properly in days I am exhausted",
"I haven't slept more than three hours in a week",
"sleep deprived and stressed I can barely function",
"I feel like I am drowning in responsibilities",
"the stress has been building for weeks and I am at my limit",
"I have back to back deadlines and no time to breathe",
]
lookup = {text: label for text, label in _CONTRAST_SAMPLES}
for text in texts_that_must_be_stressed:
assert text in lookup, (
f"Expected stressed example missing from _CONTRAST_SAMPLES: {text!r}"
)
assert lookup[text] == 1, (
f"Example should be label=1 but got {lookup[text]}: {text!r}"
)
def test_positive_resolution_examples_are_non_stressed(self):
"""Contrast/concession examples with positive resolution must be label=0."""
texts_that_must_be_non_stressed = [
"I am tired but happy",
"I feel overwhelmed but satisfied",
"was stressed but things are better now",
"I feel calm and content today",
"I don't have any worries today",
"nothing is going wrong right now",
]
lookup = {text: label for text, label in _CONTRAST_SAMPLES}
for text in texts_that_must_be_non_stressed:
assert text in lookup, (
f"Expected non-stressed example missing from _CONTRAST_SAMPLES: {text!r}"
)
assert lookup[text] == 0, (
f"Example should be label=0 but got {lookup[text]}: {text!r}"
)
def test_all_labels_are_binary(self):
"""Every entry in _CONTRAST_SAMPLES must have label 0 or 1."""
for text, label in _CONTRAST_SAMPLES:
assert label in (0, 1), (
f"Invalid label {label!r} for sample: {text!r}"
)
def test_both_classes_represented(self):
"""_CONTRAST_SAMPLES must contain at least one example of each class."""
labels = {label for _, label in _CONTRAST_SAMPLES}
assert 0 in labels, "_CONTRAST_SAMPLES has no label=0 (non-stressed) examples"
assert 1 in labels, "_CONTRAST_SAMPLES has no label=1 (stressed) examples"
|