CoralSRT: Revisiting Coral Reef Semantic Segmentation by Feature
Rectification via Self-supervised Guidance
Ziqiang Zheng1
Yuk-Kwan Wong1
Binh-Son Hua2
Jianbo Shi3
Sai-Kit Yeung1
1The Hong Kong University of Science and Technology
2Trinity College Dublin
3University of Pennsylvania
International Conference on Computer Vision, ICCV 2025
Abstract
We investigate coral reef semantic segmentation, in which
multifaceted factors, like genes, environmental changes, and
internal interactions, can lead to highly unpredictable growth
patterns. Existing segmentation approaches in both computer
vision and coral reef communities have failed to incorporate the
intrinsic properties of corals, specifically their
self-repeated, asymmetric, and amorphous distribution of
elements, into model design. We propose
CoralSRT, a feature
rectification module via self-supervised guidance, to reduce the
stochasticity of coral features extracted by pretrained
foundation models (FMs), as demonstrated in
Fig. 1. Our insight is that while different corals are highly
dissimilar, individual corals within the same growth exhibit
strong self-affinity. Using a superset of features from FMs
learned by various pretext tasks, we extract a pattern related
to the intrinsic properties of each coral to strengthen
within-segment affinity, aligning with centrality. We
investigate features from FMs that were optimized by various
pretext tasks on significantly large-scale unlabeled or labeled
data, which already contain rich information for modeling both
within-segment and cross-segment affinities, enabling the
adaptation of FMs for coral segmentation. CoralSRT can rectify
features from FMs to more efficient features for label
propagation and lead to further significant semantic
segmentation performance gains, all without requiring additional
human supervision, retraining/finetuning FMs or even
domain-specific data. These advantages help reduce human effort
and the need for domain expertise in data collection and
labeling. Our method is easy to implement, and also task- and
model-agnostic. CoralSRT bridges the self-supervised
pre-training and supervised training in the feature space, also
offering insights for segmenting elements/stuffs (e.g., grass, plants, cells, and biofoulings).
Qualitative Results
Fig. 3
The sparse-to-dense conversion results of various algorithms
using 100 randomly sampled labeled sparse points.
Fig. 4
PCA (first 3 components) visualization of features. FeatUp,
DVT and CoralSRT are using DINOv2 features.
Fig. 5
PCA visualization of original, rectified (median value based
and without training), and CoralSRT features.
© 2025 Ziqiang Zheng, Yuk-Kwan Wong, Binh-Son Hua, Jianbo Shi,
Sai-Kit Yeung. All rights reserved.