image imagewidth (px) 2k 2k |
|---|
Dataset Description
Overview
- The dataset is organized into three roots:
good_weld: labeled non-defect reference runs (750runs in43configuration folders).defect_data_weld: labeled defect runs (1580runs in80configuration folders).test_data: anonymized evaluation set (90samples namedsample_0001...sample_0090).- Training-available labeled pool (
good_weld+defect_data_weld) contains2330runs. - Each run/sample is multimodal and typically includes:
- one sensor CSV, one FLAC audio file, one AVI video file
Weld Configuration Metadata
- Weld setup/context is encoded in labeled folder names (for example: joint type such as
buttorplane_plate, material tags such asFe410orBSK46, and date-like tokens). - Run IDs follow a pattern like
04-03-23-0010-11, where the final two-digit suffix (11) is the defect/quality code. - For fair hackathon evaluation,
test_datahas anonymized folder names (sample_XXXX) and neutral filenames (sensor.csv,weld.flac,weld.avi) to reduce label leakage. - Ground-truth linkage for evaluation is kept separately in
test_data_ground_truth.csv.
Labels and Defect Definitions
| Code | Label | Practical definition | Train pool count (good_weld+defect_data_weld) |
|---|---|---|---|
00 |
good_weld |
Acceptable weld (no target defect) | 750 |
01 |
excessive_penetration |
Over-penetration through joint root | 479 |
02 |
burn_through |
Severe over-penetration causing burn-through/hole | 317 |
06 |
overlap |
Weld metal overlap without proper fusion at toe/root region | 155 |
07 |
lack_of_fusion |
Incomplete fusion between weld metal and base material | 320 |
08 |
excessive_convexity |
Excessively convex weld bead profile | 159 |
11 |
crater_cracks |
Cracks near weld crater/termination zone | 150 |
Folder Structure
good_weld/
<configuration_folder>/
<run_id>/
<run_id>.csv
<run_id>.flac
<run_id>.avi
images/*.jpg
defect_data_weld/
<defect_configuration_folder>/
<run_id>/
<run_id>.csv
<run_id>.flac
<run_id>.avi
images/*.jpg
test_data/
sample_0001/
sensor.csv
weld.flac
weld.avi
images/*.jpg
...
sample_0090/
- Evaluation helper files:
test_data_manifest.csv: file paths for each anonymized sample.test_data_ground_truth.csv: mapping fromsample_idto true label (for evaluator use).
Sensor Features (Number of Features)
- Original labeled CSV schema has
10columns: Date,Time,Part No,Pressure,CO2 Weld Flow,Feed,Primary Weld Current,Wire Consumed,Secondary Weld Voltage,Remarks.- Core process/sensor channels used for modeling are typically
6numeric features: Pressure,CO2 Weld Flow,Feed,Primary Weld Current,Wire Consumed,Secondary Weld Voltage.test_data/sensor.csvintentionally removesPart Noto prevent ID leakage, so evaluation CSVs have9columns.
Concise and Informative Guideline
- Use
good_weldanddefect_data_weldonly for training and validation. - Split by run/group (not random rows) to avoid leakage across near-duplicate temporal segments.
- Treat each run/sample as multimodal (
sensor.csv+ audio + video + images), then build binary (defectvsgood) and multi-class (defect type) models. - Run final inference only on anonymized
test_dataand export predictions keyed bysample_id. - Keep
test_data_ground_truth.csvstrictly for organizer-side scoring or final offline evaluation after predictions are frozen. - Report both performance and confidence quality (for example: F1, Macro-F1, calibration/ECE) and include failure-case examples.
Phase 1: Data preparation + dataset + dashboard + overall analysis + feature engineering
Goal: produce a clean, reproducible dataset + an analysis dashboard that explains what’s in the data and what signals might matter.
What they must do
Ingest the dataset
- Load video + audio + labels/metadata.
- Validate files: missing, corrupt, mismatched IDs, inconsistent durations.
Define the unit of prediction
- Decide what one “sample” means (whole weld, fixed-length segment, windowed chunks, etc.).
- Ensure labels align to that unit.
Create a reproducible split
- Train/validation/test split that avoids leakage (split by session/part/run if applicable).
- Save split files so results are repeatable.
Preprocess and standardize
- Make audio/video consistent (sampling rate/FPS, resizing, normalization, trimming/padding policy).
- Handle variable length (padding, cropping, pooling, sliding windows).
Feature engineering (optional, but if used it must be documented)
- Produce derived signals/features from audio/video/metadata (any representation is fine).
- Keep a clear mapping from raw inputs → engineered inputs.
Dashboard (must show)
- Dataset overview: counts, durations, missing/corrupt stats.
- Label distributions: defect vs non-defect, defect-type counts.
- Representative examples: video preview + audio preview (waveform/spectrogram or equivalent).
- Basic data quality indicators: class imbalance, outliers, noise, sync issues (if relevant).
- Exportable reports: ability to save plots/tables or generate a summary.
Phase 1 output package
dataset/or loader pipeline that can recreate the dataset- split definition files
- dashboard app/notebook
- short “data card” summary (1 page) describing assumptions and preprocessing choices
Phase 2: Defect detection (binary classification) with confidence
Goal: build a model that outputs defect vs non-defect plus a confidence score for each prediction.
What they must do
Train a binary classifier
- Input: audio/video (and any engineered features) per sample.
- Output: probability/score for “defect”.
Produce confidence
- Define what confidence means (typically a calibrated probability).
- Confidence must be reported per prediction.
Set a decision rule
- Thresholding policy to convert score → defect/non-defect.
- Threshold must be fixed for test-time scoring (not adjusted after seeing test labels).
Evaluate on validation
- Report core binary metrics (listed below).
- Show error breakdown (false positives/false negatives) and examples.
Create an inference pipeline
- Script that takes the test split and writes predictions in the required format.
Phase 2 output package
trained model checkpoint(s)
inference script (one command run)
predictions_binary.csv(or combined file) with:sample_id,p_defect,pred_defect,confidence
evaluation report/plots in the dashboard
Phase 3: Defect type classification (multi-class)
Goal: if a weld is defective, predict which defect type, with confidence.
What they must do
Train a defect-type classifier
- Input: same sample representation.
- Output: defect type probabilities (or scores).
Define handling of non-defect samples
Either:
- classify defect type only when defect is predicted/known, OR
- include “none” as a class.
Whichever they choose, it must match the evaluation spec and be consistent.
Report confidence for type
- Provide a confidence score for the chosen defect type (top-1 probability or calibrated).
Evaluate
- Report multi-class metrics (listed below), especially per-class results due to imbalance.
Integrate with Phase 2
- Final output should be coherent: non-defect → type is “none”; defect → type predicted.
Phase 3 output package
model checkpoint(s)
inference script producing:
pred_defect_type,p_type_*(optional),type_confidence
evaluation report (per-type performance)
Evaluation criteria
A) Model metrics (primary)
Submission CSV (required)
Teams must submit one CSV file with this exact schema:
sample_id,pred_label_code,p_defect
sample_0001,11,0.94
sample_0002,00,0.08
...
sample_0090,06,0.81
Submission rules:
- Exactly
90rows (one row per sample intest_data_manifest.csv) sample_idmust match exactly (sample_0001...sample_0090), with no duplicatespred_label_codemust be one of:00,01,02,06,07,08,11p_defectmust be numeric in[0,1]
Scoring interpretation:
- Binary prediction is derived as:
pred_defect = (pred_label_code != "00") - Type prediction is the submitted
pred_label_code
1) Defect vs non-defect (binary)
Use these as the core:
- F1 (Defect as positive class)
- Precision / Recall (Defect)
- ROC-AUC
- PR-AUC
- Confusion matrix counts (TP/FP/FN/TN)
2) Defect type (multi-class)
Use these:
- Macro F1 (treats each defect type equally, good for imbalance)
- Per-class Precision/Recall/F1
- Weighted F1 (secondary)
single final score:
FinalScore = 0.6 * Binary_F1 + 0.4 * Type_MacroF1
B) Engineering & product quality (secondary)
UI / Dashboard (clean and usable)
- Clear navigation, readable plots/tables, consistent labels
- Shows the required dataset stats + evaluation views
- Fast enough to use during a demo (no 5-minute refreshes)
Clean code & reproducibility
- One-command run for training/inference
- Clear folder structure, requirements/environment file
- No hardcoded paths, no mystery constants without comments
- Reproducible splits + fixed random seeds (where relevant)
Presentation & explanation
Clear statement of:
- sample definition
- preprocessing assumptions
- model outputs and how confidence is computed
- strengths/weaknesses and common failure cases
Demo includes: dashboard + a few correctly/incorrectly predicted examples
- Downloads last month
- 40