UNet-eVAE: Iterative Refinement Using VAE Embodied Learning for Endoscopic Image Segmentation

Soumya Gupta* (Corresponding Author), Sharib Ali, Xu Zhang, Binod Bhattarai, Jens Rittscher

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingPublished conference contribution

Abstract

While endoscopy is routinely used for surveillance, high operator dependence demands robust automated image analysis methods. Automated segmentation of region-of-interest (ROI) that includes lesions, inflammations, and instruments can serve to cope with the operator dependence problem in this field. Most supervised methods are developed by fitting models on the available ground truth mask samples only. This work proposes a joint training approach using the UNet coupled with a variational auto-encoder (VAE) to improve endoscopic image segmentation by exploiting original samples, predicted masks and ground truth masks. In the proposed UNet-eVAE, VAE utilises the masks to constrain ROI-specific feature representations for reconstruction as an auxiliary task. The fine-grained spatial information from VAE is fused with the UNet decoder to enrich the feature representations and improve segmentation performance. Our experimental results on both colonoscopy and ureteroscopy datasets demonstrate that the proposed architecture can learn robust representations and generalise segmentation performance on unseen samples while improving the baseline.
Original languageEnglish
Title of host publicationMachine Learning in Medical Imaging
Subtitle of host publication13th International Workshop, MLMI 2022, Held in Conjunction with MICCAI 2022, Singapore, September 18, 2022, Proceedings
PublisherSpringer Nature
Pages161-170
Number of pages10
ISBN (Electronic)978-3-031-21014-3
ISBN (Print)978-3-031-21013-6
DOIs
Publication statusPublished - 16 Dec 2022

Publication series

NameLecture Notes in Computer Science
PublisherSpringer Nature
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Bibliographical note

We would like to thank Boston Scientific for funding this project (Grant No: DFR04690). SG and BT are funded by BSC, BB is funded by EndoMapper Horizon 2020 FET (GA 863146), SA and JR were supported by the NIHR Oxford Biomedical Research Centre.

Fingerprint

Dive into the research topics of 'UNet-eVAE: Iterative Refinement Using VAE Embodied Learning for Endoscopic Image Segmentation'. Together they form a unique fingerprint.

Cite this