Skip to content

[TMI 2026] Spatio-Temporal Representation Decoupling and Enhancement for Federated Instrument Segmentation in Surgical Videos

Notifications You must be signed in to change notification settings

jinlab-imvr/FedST

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

FedST: Spatio-Temporal Representation Decoupling and Enhancement for Federated Instrument Segmentation in Surgical Videos

Official implementation of the paper "Spatio-Temporal Representation Decoupling and Enhancement for Federated Instrument Segmentation in Surgical Videos" (arXiv:2506.23759)

Illustration of the multi-site federated surgical segmentation challenge.

In this paper, we propose a novel Personalized FL scheme, **Spatio-Temporal Representation Decoupling and Enhancement (FedST)**, which wisely leverages surgical domain knowledge during both local-site and global-server training to boost segmentation.

⚙️ Environment Setup

Ensure the following dependencies are installed:

  • Python ≥ 3.8
  • PyTorch ≥ 1.12
  • CUDA ≥ 11.3

Example installation

conda create -n pfedsis python=3.9
conda activate pfedsis
conda install pytorch==2.5.1 torchvision==0.20.1 torchaudio==2.5.1 pytorch-cuda=11.8 -c pytorch -c nvidia

cd FedST
pip install -r requirement.txt

📦 Dataset Preparation

  1. Download: Download the FL_Dataset.zip file from the Files tab.
  2. Unzip: Unzip the file to your project directory.

⚠️ Note on Data Format: The original codebase relies on .npy files for accelerated data loading. However, to optimize download speeds and storage on Hugging Face, the dataset is provided in .png format.

To resolve this, please choose one of the following:

  • Convert the data: Pre-process the .png images back into .npy format.
  • Update the loader: Modify dataloaders/robotics_dataloader.py to load .png files directly instead of .npy.
# Example command
unzip FL_Dataset.zip

## 🚀 Training and Evaluation

You can start the full training with:

```bash
python3 final_method.py --exp <save_path> --max_epoch 300 --dataset robotics

--exp specifies the output directory to save checkpoints and logs.


🧠 Acknowledgements

This research was supported by collaborative efforts within the iMVR Lab and multiple surgical data initiatives. We thank the contributors of the public surgical video datasets that made this benchmark possible.

📧 Correspondence

For any inquiries regarding this project, please contact Zheng Fang.

About

[TMI 2026] Spatio-Temporal Representation Decoupling and Enhancement for Federated Instrument Segmentation in Surgical Videos

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages