Target-aware cross-modality unsupervised domain adaptation for vestibular schwannoma and cochlea segmentation

Sci Rep. 2024 Nov 13;14(1):27883. doi: 10.1038/s41598-024-77633-x.

Abstract

There is growing interest in research on segmentation for the vestibular schwannoma (VS) and cochlea using high-resolution T2 (hrT2) imaging over contrast-enhanced T1 (ceT1) imaging due to the contrast agent side effects. However, the hrT2 imaging remains a problem of insufficient annotated data, which is fatal for building more robust segmentation models. To address the issue, recent studies have adopted unsupervised domain adaptation approaches that translate ceT1 images to hrT2 images. However, previous studies did not consider the size and visual characteristics of the target objects, such as VS and cochlea, during image translation. Specifically, those works simply performed normalization on the entire image without considering its significant impact on the quality of the translated images. These approaches tend to erase the small target objects, making it difficult to preserve the structure of these objects when generating pseudo-target images. Furthermore, they may also struggle to accurately reflect the unique style of the target objects within the images. Therefore, we propose a target-aware unsupervised domain adaptation framework, designed for translating target objects, each tailored to their unique visual characteristics and size using target-aware normalization. We demonstrate the superiority of the proposed framework on a publicly available challenge dataset. Codes are available at https://github.com/Bokyeong-Kang/TANQ .

MeSH terms

  • Cochlea* / diagnostic imaging
  • Cochlea* / pathology
  • Contrast Media
  • Humans
  • Image Processing, Computer-Assisted / methods
  • Magnetic Resonance Imaging* / methods
  • Neuroma, Acoustic* / diagnostic imaging
  • Neuroma, Acoustic* / pathology

Substances

  • Contrast Media