RunningPie's picture
Upload folder using huggingface_hub
97e3fb6 verified
---
language:
- en
tags:
- sign language recognition
- emergency response
- computer vision
---
# CLARIS - Critical Emergency Sign Language Dataset
This dataset is a curated subset of the "Google - Isolated Sign Language Recognition" dataset, specifically filtered for the **CLARIS (Clear and Live Automated Response for Inclusive Safety)** project.
## Dataset Description
The primary goal of the CLARIS project is to develop a mobile application that provides a lifeline for the Deaf community during emergencies. This dataset was created to train a proof-of-concept AI model capable of recognizing a vocabulary of critical emergency-related signs.
The data consists of pre-extracted landmark coordinates from video clips of isolated signs. It originates from the [Google - Isolated Sign Language Recognition Kaggle Competition](https://www.kaggle.com/competitions/asl-signs).
## Dataset Structure
The dataset is provided in both CSV and Parquet (coming soon) formats. Each row represents the coordinates of a single landmark in a single frame of a video sequence.
| Column | Dtype | Description |
| ---------------- | ------- | --------------------------------------------------------------------------- |
| `frame` | int16 | The frame number within the sequence. |
| `row_id` | object | A unique identifier for the landmark within the frame. |
| `type` | object | The type of landmark (`face`, `left_hand`, `right_hand`, `pose`). |
| `landmark_index` | int16 | The index of the landmark within its type. |
| `x` | float64 | The normalized x-coordinate of the landmark. |
| `y` | float64 | The normalized y-coordinate of the landmark. |
| `z` | float64 | The normalized z-coordinate of the landmark (depth). |
| `path` | object | The path to the original source parquet file for the sequence. |
| `participant_id` | int64 | A unique identifier for the participant (signer). |
| `sequence_id` | int64 | A unique identifier for the sign sequence. |
| `sign` | object | The ground truth label for the sign being performed. |
## Curation Process
To create a focused dataset for our specific use case, we performed a two-step curation process:
1. **Vocabulary Filtering:** We selected **62 signs** deemed most relevant for describing medical, fire, or intruder emergencies.
2. **Participant Filtering:** To create a manageable dataset for rapid prototyping, we constrained the data to sequences from **two distinct participants** who had a balanced distribution of the target signs.
This process resulted in a final dataset containing **1,719 unique sign sequences**, comprising over 37 million landmark rows.
## Usage
We recommend using the Parquet file for faster loading times.
```python
import pandas as pd
# Load the full curated dataset
df = pd.read_parquet('claris_curated_dataset.parquet')
# Or load the smaller, subsampled version
df_sample = pd.read_parquet('claris_subsample_dataset.parquet')
print(df.head())
```
## Link to Project Notebook
The complete methodology, including data preprocessing, model training, and analysis, can be found in our Kaggle notebook:
https://www.kaggle.com/code/eveelyn/datathon2025-med