FetalCLIP / README.md
Numan Saeed
Add Not Sure feedback, save preprocessed scans, add Help tab
f84688d
metadata
title: FetalCLIP
emoji: ๐Ÿ‘ถ
colorFrom: green
colorTo: blue
sdk: docker
pinned: false
license: apache-2.0

FetalCLIP - Fetal Ultrasound Analysis

Foundation Model for Zero-Shot Fetal Ultrasound Analysis

Features

  • ๐Ÿ”ฌ View Classification: Classify ultrasound images into 13 anatomical views
  • ๐Ÿ“… Gestational Age Estimation: Estimate gestational age from fetal brain ultrasounds
  • ๐Ÿฅ DICOM Support: Full preprocessing pipeline for medical DICOM files
  • ๐Ÿ–ผ๏ธ PNG/JPEG Support: Basic preprocessing for standard image files

How to Use

  1. Upload a fetal ultrasound image (PNG, JPEG, or DICOM)
  2. Click "Classify View" to identify the anatomical plane
  3. View the top predictions with confidence scores

Model

This demo uses the FetalCLIP model, a vision-language foundation model trained on fetal ultrasound images.

  • Model: numansaeed/fetalclip-model
  • Architecture: ViT-L/14 based CLIP model
  • Training: Contrastive learning on fetal ultrasound-text pairs

Supported Views

  1. Fetal abdomen
  2. Fetal brain (transventricular)
  3. Fetal brain (transthalamic)
  4. Fetal brain (transcerebellar)
  5. Fetal femur
  6. Fetal heart (4-chamber)
  7. Fetal heart (LVOT)
  8. Fetal heart (RVOT)
  9. Fetal heart (3VV)
  10. Fetal kidney
  11. Fetal face (lips)
  12. Fetal spine (coronal)
  13. Fetal spine (sagittal)

Citation

If you use this model, please cite:

@article{maani2025fetalclip,
  title={FetalCLIP: A Visual-Language Foundation Model for Fetal Ultrasound Image Analysis},
  author={Maani, Fadillah and Saeed, Numan and Saleem, Tausifa and Farooq, Zaid and Alasmawi, Hussain and Diehl, Werner and Mohammad, Ameera and Waring, Gareth and Valappi, Saudabi and Bricker, Leanne and Yaqub, Mohammad},
  journal={arXiv preprint arXiv:2502.14807},
  year={2025}
}

Links