SDSimPoint: Shallow-Deep Similarity Learning for Few-Shot Point Cloud Semantic Segmentation

Page view(s)
0
Checked on
SDSimPoint: Shallow-Deep Similarity Learning for Few-Shot Point Cloud Semantic Segmentation
Title:
SDSimPoint: Shallow-Deep Similarity Learning for Few-Shot Point Cloud Semantic Segmentation
Journal Title:
IEEE Transactions on Neural Networks and Learning Systems
Keywords:
Publication Date:
24 March 2025
Citation:
Wang, J., Zhu, H., Guo, H., Al Mamun, A., Xiang, C., de Silva, C. W., & Heng Lee, T. (2025). SDSimPoint: Shallow-Deep Similarity Learning for Few-Shot Point Cloud Semantic Segmentation. IEEE Transactions on Neural Networks and Learning Systems, 36(6), 10043–10056. https://doi.org/10.1109/tnnls.2025.3543620
Abstract:
Three-dimensional point cloud semantic segmentation is a fundamental task in computer vision. As the fully supervised approaches suffer from the generalization issue with limited data, few-shot point cloud segmentation models have been proposed to address the flexible adaptation. Nevertheless, due to the class-agnostic nature of the few-shot pretraining, its pretrained feature extractor is hard to capture the class-related intrinsic and abstract information. Therefore, we introduce the new concept of shallow and deep similarities and propose a shallow-deep similarity learning network (SDSimPoint) that aims to learn both shallow (superficial geometry, color, etc.) and deep similarities (intrinsic context and semantics, etc.) between the support and query samples, thereby boosting the performance. Moreover, we design a beyond-episode attention module (BEAM) to enlarge the region of the attention mechanism from a single episode to the entire dataset by utilizing the memory units, which enhances the extraction ability to better capture the shallow and deep similarities. Furthermore, our distance metric function is learnable in the proposed framework, which can better adapt to complex data distributions. Our proposed SDSimPoint consistently demonstrates substantial improvements compared to baseline approaches across various datasets in diverse few-shot point cloud semantic segmentation settings.
License type:
Publisher Copyright
Funding Info:
This research / project is supported by the Agency for Science, Technology and Research - RIE2025 Industry Alignment Fund - Pre-Positioning Advanced ROS2-native Platform Technologies for Cross sectorial Robotics Adoption
Grant Reference no. : M21K1a0104
Description:
© 2025 IEEE.  Personal use of this material is permitted.  Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works
ISSN:
2162-237X
2162-2388
Files uploaded:

File Size Format Action
tnnls-2024-p-33109-r1-finalversion.pdf 35.67 MB PDF Open