File size: 3,681 Bytes
e02ee6b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
97e3fb6
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
---

language:
  - en
tags:
  - sign language recognition
  - emergency response
  - computer vision
---


# CLARIS - Critical Emergency Sign Language Dataset

This dataset is a curated subset of the "Google - Isolated Sign Language Recognition" dataset, specifically filtered for the **CLARIS (Clear and Live Automated Response for Inclusive Safety)** project.

## Dataset Description

The primary goal of the CLARIS project is to develop a mobile application that provides a lifeline for the Deaf community during emergencies. This dataset was created to train a proof-of-concept AI model capable of recognizing a vocabulary of critical emergency-related signs.

The data consists of pre-extracted landmark coordinates from video clips of isolated signs. It originates from the [Google - Isolated Sign Language Recognition Kaggle Competition](https://www.kaggle.com/competitions/asl-signs).

## Dataset Structure

The dataset is provided in both CSV and Parquet (coming soon) formats. Each row represents the coordinates of a single landmark in a single frame of a video sequence.

| Column           | Dtype   | Description                                                                 |
| ---------------- | ------- | --------------------------------------------------------------------------- |
| `frame`          | int16   | The frame number within the sequence.                                       |
| `row_id`         | object  | A unique identifier for the landmark within the frame.                      |
| `type`           | object  | The type of landmark (`face`, `left_hand`, `right_hand`, `pose`).             |
| `landmark_index` | int16   | The index of the landmark within its type.                                  |
| `x`              | float64 | The normalized x-coordinate of the landmark.                                |
| `y`              | float64 | The normalized y-coordinate of the landmark.                                |
| `z`              | float64 | The normalized z-coordinate of the landmark (depth).                        |
| `path`           | object  | The path to the original source parquet file for the sequence.              |
| `participant_id` | int64   | A unique identifier for the participant (signer).                           |
| `sequence_id`    | int64   | A unique identifier for the sign sequence.                                  |
| `sign`           | object  | The ground truth label for the sign being performed.                        |

## Curation Process

To create a focused dataset for our specific use case, we performed a two-step curation process:

1.  **Vocabulary Filtering:** We selected **62 signs** deemed most relevant for describing medical, fire, or intruder emergencies.
2.  **Participant Filtering:** To create a manageable dataset for rapid prototyping, we constrained the data to sequences from **two distinct participants** who had a balanced distribution of the target signs.

This process resulted in a final dataset containing **1,719 unique sign sequences**, comprising over 37 million landmark rows.

## Usage

We recommend using the Parquet file for faster loading times.

```python

import pandas as pd



# Load the full curated dataset

df = pd.read_parquet('claris_curated_dataset.parquet')



# Or load the smaller, subsampled version

df_sample = pd.read_parquet('claris_subsample_dataset.parquet')



print(df.head())

```

## Link to Project Notebook

The complete methodology, including data preprocessing, model training, and analysis, can be found in our Kaggle notebook:
https://www.kaggle.com/code/eveelyn/datathon2025-med