|
|
--- |
|
|
license: mit |
|
|
dataset_info: |
|
|
features: |
|
|
- name: subreddit |
|
|
dtype: string |
|
|
- name: created_at |
|
|
dtype: timestamp[ns, tz=US/Central] |
|
|
- name: retrieved_at |
|
|
dtype: timestamp[ns, tz=US/Central] |
|
|
- name: type |
|
|
dtype: string |
|
|
- name: text |
|
|
dtype: string |
|
|
- name: score |
|
|
dtype: int64 |
|
|
- name: post_id |
|
|
dtype: string |
|
|
- name: parent_id |
|
|
dtype: string |
|
|
--- |
|
|
|
|
|
# Top Reddit Posts Daily |
|
|
|
|
|
## Dataset Summary |
|
|
|
|
|
A continuously-updated snapshot of public Reddit discourse on AI news. Each night a GitHub Actions cron job |
|
|
|
|
|
1. **Scrapes** new submissions from a configurable list of subreddits (→ `data_raw/`) |
|
|
2. **Classifies** each post with a DistilBERT sentiment model served on Replicate (→ `data_scored/`) |
|
|
3. **Summarises** daily trends for lightweight front-end consumption (→ `daily_summary/`) |
|
|
|
|
|
The result is an easy-to-query, time-stamped record of Reddit sentiment that can be used for NLP research, social-media trend analysis, or as a teaching dataset for end-to-end MLOps. |
|
|
|
|
|
Source code https://github.com/halstonblim/reddit_sentiment_pipeline |
|
|
|
|
|
Currently configured to scrape only the top daily posts and comments to respect rate limits |
|
|
``` |
|
|
subreddits: |
|
|
- name: artificial |
|
|
post_limit: 100 |
|
|
comment_limit: 10 |
|
|
- name: LocalLLaMA |
|
|
post_limit: 100 |
|
|
comment_limit: 10 |
|
|
- name: singularity |
|
|
post_limit: 100 |
|
|
comment_limit: 10 |
|
|
- name: OpenAI |
|
|
post_limit: 100 |
|
|
comment_limit: 10 |
|
|
``` |
|
|
|
|
|
## Supported Tasks |
|
|
This dataset can be used for: |
|
|
- Text classification (e.g., sentiment analysis) |
|
|
- Topic modeling |
|
|
- Language generation and summarization |
|
|
- Time‑series analysis of Reddit activity |
|
|
|
|
|
|
|
|
## Languages |
|
|
- English, no filtering is currently done on the raw text |
|
|
|
|
|
## Dataset Structure |
|
|
|
|
|
``` |
|
|
hblim/top_reddit_posts_daily/ |
|
|
└── data_raw/ # contains raw data scraped from reddit |
|
|
├── 2025‑05‑01.parquet |
|
|
├── 2025‑05‑01.parquet |
|
|
└── … |
|
|
└── data_scored/ # contains same rows as raw data but with sentiment scores |
|
|
├── 2025‑05‑01.parquet |
|
|
├── 2025‑05‑01.parquet |
|
|
└── … |
|
|
└── subreddit_daily_summary.csv/ # contains daily summaries of sentiment averages grouped by (day, subreddit) |
|
|
``` |
|
|
|
|
|
### Data Fields |
|
|
|
|
|
| Name | Type | Description | |
|
|
|----------------|------------|------------------------------------------------------------------------------| |
|
|
| `subreddit` | `string` | Name of the subreddit (e.g. “GooglePixel”) | |
|
|
| `created_at` | `datetime` | UTC timestamp when the post/comment was originally created | |
|
|
| `retrieved_at` | `datetime` | Local timezone timestamp when this data was scraped | |
|
|
| `type` | `string` | `"post"` or `"comment"` | |
|
|
| `text` | `string` | For posts: `title + "\n\n" + selftext`; for comments: comment body | |
|
|
| `score` | `int` | Reddit score (upvotes – downvotes) | |
|
|
| `post_id` | `string` | Unique Reddit ID for the post or comment | |
|
|
| `parent_id` | `string` | For comments: the parent comment/post ID; `null` for top‑level posts | |
|
|
|
|
|
Example entry: |
|
|
|
|
|
| Field | Value | |
|
|
|--------------|-----------------------------------------------------------| |
|
|
| subreddit | apple | |
|
|
| created_at | 2025-04-17 19:59:44-05:00 | |
|
|
| retrieved_at | 2025-04-18 12:46:10.631577-05:00 | |
|
|
| type | post | |
|
|
| text | Apple wanted people to vibe code Vision Pro apps with Siri | |
|
|
| score | 427 | |
|
|
| post_id | 1k1sn9w | |
|
|
| parent_id | None | |
|
|
|
|
|
|
|
|
## Data Splits |
|
|
There are no explicit train/test splits. Data is organized by date under the `data_raw/` or `data_scored/` folder. |
|
|
|
|
|
## Dataset Creation |
|
|
|
|
|
- A Python script (`scrape.py`) runs daily, fetching the top N posts and top M comments per subreddit. |
|
|
- Posts are retrieved via PRAW’s `subreddit.top(time_filter="day")`. |
|
|
- Data is de‑duplicated against the previous day’s `post_id` values. |
|
|
- Stored as Parquet under `data_raw/{YYYY‑MM‑DD}.parquet`. |
|
|
|
|
|
## License |
|
|
This dataset is released under the [MIT License](https://opensource.org/licenses/MIT). |
|
|
|
|
|
## Citation |
|
|
If you use this dataset, please cite it as: |
|
|
|
|
|
```bibtex |
|
|
@misc{lim_top_reddit_posts_daily_2025, |
|
|
title = {Top Reddit Posts Daily: Scraped Daily Top Posts and Comments from Subreddits}, |
|
|
author = {Halston Lim}, |
|
|
year = {2025}, |
|
|
publisher = {Hugging Face Datasets}, |
|
|
howpublished = {\url{https://huggingface.co/datasets/hblim/top_reddit_posts_daily}} |
|
|
} |
|
|
``` |
|
|
|
|
|
## Usage Example |
|
|
|
|
|
### Example A: Download and load a single day via HF Hub |
|
|
|
|
|
```python |
|
|
from huggingface_hub import HfApi |
|
|
import pandas as pd |
|
|
|
|
|
api = HfApi() |
|
|
repo_id = "hblim/top_reddit_posts_daily" |
|
|
|
|
|
date_str = "2025-04-18" |
|
|
today_path = api.hf_hub_download( |
|
|
repo_id=repo_id, |
|
|
filename=f"data_raw/{date_str}.parquet", |
|
|
repo_type="dataset" |
|
|
) |
|
|
df_today = pd.read_parquet(today_path) |
|
|
print(f"Records for {date_str}:") |
|
|
print(df_today.head()) |
|
|
``` |
|
|
|
|
|
### Example B: List, download, and concatenate all days |
|
|
|
|
|
```python |
|
|
from huggingface_hub import HfApi |
|
|
import pandas as pd |
|
|
|
|
|
api = HfApi() |
|
|
repo_id = "hblim/top_reddit_posts_daily" |
|
|
|
|
|
# 1. List all parquet files in the dataset repo |
|
|
all_files = api.list_repo_files(repo_id, repo_type="dataset") |
|
|
parquet_files = sorted([f for f in all_files if f.startswith("data_raw/") and f.endswith(".parquet")]) |
|
|
|
|
|
# 2. Download each shard and load with pandas |
|
|
dfs = [] |
|
|
for shard in parquet_files: |
|
|
local_path = api.hf_hub_download(repo_id=repo_id, filename=shard, repo_type="dataset") |
|
|
dfs.append(pd.read_parquet(local_path)) |
|
|
|
|
|
# 3. Concatenate into one DataFrame |
|
|
df_all = pd.concat(dfs, ignore_index=True) |
|
|
print(f"Total records across {len(dfs)} days: {len(df_all)}") |
|
|
``` |
|
|
|
|
|
## Limitations & Ethics |
|
|
- **Bias:** Data reflects Reddit’s user base and community norms, which may not generalize. |
|
|
- **Privacy:** Only public content is collected; no personally identifiable information is stored. |