Spaces:
Sleeping
Sleeping
File size: 1,964 Bytes
ee412eb |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 |
import streamlit as st
def show():
st.markdown('<div class="main-header">๐ฅ Video Surgical Scene Understanding Dashboard</div>', unsafe_allow_html=True)
st.markdown("---")
# Welcome and overall description
st.markdown("## Welcome to the Surgical Scene Analysis Platform")
st.markdown("""
This platform demonstrates an end-to-end pipeline for automated understanding of surgical scenes from video data.
The system leverages advanced computer vision and AI models to analyze surgical workflows, recognize tools, and generate scene-level captions.
Navigate through the sidebar to test the system, explore datasets, or learn more about the project.
""")
st.markdown("---")
st.markdown("## ๐ Pipeline Overview")
st.markdown("""
The surgical scene understanding pipeline consists of the following main steps:
1. **Frame Extraction**: Select or upload three consecutive frames from a surgical video.
2. **Segmentation**: Use the SwinUNETR model to generate a segmentation mask for the scene.
3. **Captioning**: Input the frames and mask into the MedGemma model to generate a descriptive caption or scene graph.
4. **Results & Analysis**: Review the generated mask and caption to understand the surgical context.
""")
st.markdown("---")
st.markdown("## ๐ Project Description")
st.markdown("""
This project was developed by **Team SATOR** for the ACVSS 2025 Hackathon.
Our goal is to provide an accessible, interactive demonstration of state-of-the-art surgical scene understanding using deep learning.
- **Frontend**: Streamlit Dashboard
- **Backend**: Python, PyTorch, MONAI, HuggingFace Transformers
- **Models**: SwinUNETR (segmentation), MedGemma (captioning)
- **Dataset**: MM-OR (Multimodal Operating Room)
""")
st.markdown("---")
st.info("Use the sidebar to start testing the system or to learn more about the dataset and team.")
|