Skip to content

Code for MICCAI2023: YONA: You Only Need One Adjacent Reference-frame for Accurate and Fast Video Polyp Detection

Notifications You must be signed in to change notification settings

yuncheng97/YONA

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

by Yuncheng Jiang, Zixun Zhang, Ruimao Zhang, Guanbin Li, Shuguang Cui, and Zhen Li

Introduction

framework Accurate polyp detection is essential for assisting clinical rectal cancer diagnoses. Colonoscopy videos contain richer information than still images, making them a valuable resource for deep learning methods. However, unlike common fixed-camera video, the camera-moving scene in colonoscopy videos can cause rapid video jitters, leading to unstable training for existing video detection models. In this paper, we propose the YONA (You Only Need one Adjacent Reference-frame) method, an efficient end-to-end training framework for video polyp detection. YONA fully exploits the information of one previous adjacent frame and conducts polyp detection on the current frame without multi-frame collaborations. Specifically, for the foreground, YONA adaptively aligns the current frame’s channel activation patterns with its adjacent refer- ence frames according to their foreground similarity. For the background, YONA conducts background dynamic alignment guided by inter-frame difference to eliminate the invalid features produced by drastic spatial jitters. Moreover, YONA applies cross-frame contrastive learning dur- ing training, leveraging the ground truth bounding box to improve the model’s perception of polyp and background. Quantitative and quali- tative experiments on three public challenging benchmarks demonstrate that our proposed YONA outperforms previous state-of-the-art competi- tors by a large margin in both accuracy and speed.

Clone repository

git clone https://github.com/yuncheng97/YONA.git
cd YONA/

File tree

YONA                         
├── README.md
├── data.py
├── figures
│   └── framework.png
├── model.py
├── pretrained
├── result
├── pvtv2.py
├── resnet.py
├── train.py
├── train.sh
└── utils.py

Download dataset

The training and testing datasets come from public colonoscopy benchmarks: 'SUNSEG', 'LDPolypVideo', 'CVC-VideoClinicDB' and 'PolypGen'. Download these datasets and unzip them into data folder.

Prerequisites

Pretrained

download the pretrained backbone models and put them in the pretrained folder

Training

Set your own configuration before training. Make sure you set the right dataset path in '--data_path' and save path in '--save_path'.

    python train.py

    or

    bash train.sh

Citation

  • If you find this work is helpful, please cite our paper
@article{jiang2023yona,
  title={YONA: You Only Need One Adjacent Reference-frame for Accurate and Fast Video Polyp Detection},
  author={Jiang, Yuncheng and Zhang, Zixun and Zhang, Ruimao and Li, Guanbin and Cui, Shuguang and Li, Zhen},
  journal={Medical Image Computing and Computer Assisted Intervention -- MICCAI 2023},
  year={2023}
}

About

Code for MICCAI2023: YONA: You Only Need One Adjacent Reference-frame for Accurate and Fast Video Polyp Detection

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published