license: cc-by-nc-4.0
task_categories:
- text-classification
- image-classification
- multimodal
tags:
- web-security
- phishing-detection
- html-analysis
- screenshot
- manually-verified
dataset_info:
features:
- name: url
dtype: string
description: Website URL and folder name for resource access
- name: html_content
dtype: string
description: 'Path to HTML file (format: {url}/page.html)'
- name: screenshot_content
dtype: image
description: 'Path to screenshot image (format: {url}/screenshot.png)'
- name: label
dtype: string
description: Classification label from metadata.txt (phishing/legitimate)
splits:
- name: train
num_bytes: 1024000000
num_examples: 1000
configs:
- config_name: default
data_files:
- split: train
path: data/dataset.csv
Web Content Classification Dataset with HTML and Screenshots
Dataset
Description: This dataset contains 1,000 carefully curated examples of web content designed for phishing detection classification tasks. Each example consists of a website URL along with its complete HTML content, a visual screenshot, and a manually verified classification label.
Important Note: This entire dataset was meticulously created through manual processes. Each website was individually visited, analyzed, and verified to ensure the highest quality and accuracy of labels.
Dataset Structure
The dataset is organized with a main CSV file that references all resources using a consistent folder structure:
CSV Columns:
- url: Website URL and folder name for resource access
- html_content: Path to HTML file (
{url}/page.html) - screenshot_content: Path to screenshot image (
{url}/screenshot.png) - label: Classification label (phishing/legitimate)
Directory Structure:
dataset: files: - dataset.csv folders: - name: example-domain-1.com files: - page.html - screenshot.png - metadata.txt - name: example-domain-2.com files: - page.html - screenshot.png - metadata.txt
File Reference System
The dataset uses a consistent referencing system where each URL serves as both the website address and folder name:
- HTML Content: "{url}/page.html (e.g.,
example.com/page.html)" - Screenshot: "{url}/screenshot.png (e.g.,
example.com/screenshot.png) " - Label: Value contained in
{url}/metadata.txt
Collection Methodology
Manual Curation Process
This dataset was created through an extensive manual verification process:
- Individual Website Visits: Each of the 1,000 websites was personally visited and analyzed
- Manual Verification: Every classification label was manually assigned after careful examination
- Quality Control: Each entry was individually checked for accuracy and completeness
- Content Capture: HTML and screenshots were captured while ensuring proper rendering
Data Points Collected:
data_points:
URL: "The exact folder name and web address"
HTML content: "Full source code saved as {url}/page.html"
Visual representation: "High-quality screenshot saved as {url}/screenshot.png"
Classification label: "Value stored in {url}/metadata.txt (verified as phishing/legitimate)"