Spaces:
Sleeping
Sleeping
| import streamlit as st | |
| def show(): | |
| st.markdown('<div class="main-header">π₯ Video Surgical Scene Understanding Dashboard</div>', unsafe_allow_html=True) | |
| st.markdown("---") | |
| # Welcome and overall description | |
| st.markdown("## Welcome to the Surgical Scene Analysis Platform") | |
| st.markdown(""" | |
| This platform demonstrates an end-to-end pipeline for automated understanding of surgical scenes from video data. | |
| The system leverages advanced computer vision and AI models to analyze surgical workflows, recognize tools, and generate scene-level captions. | |
| Navigate through the sidebar to test the system, explore datasets, or learn more about the project. | |
| """) | |
| st.markdown("---") | |
| st.markdown("## π Pipeline Overview") | |
| st.markdown(""" | |
| The surgical scene understanding pipeline consists of the following main steps: | |
| 1. **Frame Extraction**: Select or upload three consecutive frames from a surgical video. | |
| 2. **Segmentation**: Use the SwinUNETR model to generate a segmentation mask for the scene. | |
| 3. **Captioning**: Input the frames and mask into the MedGemma model to generate a descriptive caption or scene graph. | |
| 4. **Results & Analysis**: Review the generated mask and caption to understand the surgical context. | |
| """) | |
| st.markdown("---") | |
| st.markdown("## π Project Description") | |
| st.markdown(""" | |
| This project was developed by **Team SATOR** for the ACVSS 2025 Hackathon. | |
| Our goal is to provide an accessible, interactive demonstration of state-of-the-art surgical scene understanding using deep learning. | |
| - **Frontend**: Streamlit Dashboard | |
| - **Backend**: Python, PyTorch, MONAI, HuggingFace Transformers | |
| - **Models**: SwinUNETR (segmentation), MedGemma (captioning) | |
| - **Dataset**: MM-OR (Multimodal Operating Room) | |
| """) | |
| st.markdown("---") | |
| st.info("Use the sidebar to start testing the system or to learn more about the dataset and team.") | |